Monash University researchers have developed a new compression algorithm based on the work of a team at the Australian National University that could dramatically reduce the amount of data it takes to store an image.
Key points:The new algorithm, called hlsl, has been developed to handle large data setsThe research was conducted by Monash PhD candidate Michael C. Dickson and his teamAt the centre of the work is a new algorithm called hflax that is able to compress large data.
It’s the first time a compression algorithm has been made using techniques from mathematics and computer science.
“The original algorithm is called the Hlsl algorithm,” Dr Dickson said.
“We’ve taken the algorithm that was used by the Hcl team from the original Hcl paper and improved it to take advantage of a new type of data that was available at the time, namely the image data.”
Dr Dickson, who is also a research fellow at the National University of Singapore, said the algorithm had a number of advantages over the HCL algorithm.”HFLAX is capable of storing large amounts of information with a very small amount of memory,” he said.
Dr Dison said the new algorithm could potentially be used to store images for commercial use, and it could also be used by large organisations to store their own data.
“This new algorithm is very scalable,” he explained.
“It’s actually very easy to implement.
We can compress large amounts, and if we do, we can store that data in this very compact, memory-efficient form.”
The team of researchers said the work was important because the data could be compressed into smaller files that could be easily transmitted to clients.
“With our algorithm, we were able to make a very simple compression algorithm for data that would be very useful to data centres, to get the same amount of information that HCL has for a very large amount of storage,” Dr Coker said.
The new compression method uses the concept of “convergence”, which refers to how a computer’s algorithms work when they are combined.
“Convergence is important when we’re dealing with big data,” Dr. Dillard said.
“In our original paper, we had a lot of examples where the HFLX algorithm was very slow because it had a very low convergence.”
“The HFL algorithm has a very high convergence, and so we decided to use this convergence technique to get a little bit faster.”
He said the team’s algorithms were also capable of handling much more data than HCL, as well as larger images.
“If you have large amounts in a dataset, we have an advantage because our algorithm is able go much further,” Dr Pinto said.
In terms of the new compression technique, Dr Cotto said he was pleased that the team had been able to use mathematical techniques in order to find the right algorithm.
He said it was important that the new technology was not just used by researchers to produce high quality images, but also to make them more efficient.
“Because this is a data compression method, we’re looking for very small files, small data sets, and to be able to do that, you need to be very efficient in terms of what you’re doing with that data,” he told ABC Radio Melbourne.
“So it is a very good use of mathematics, and the mathematics are not just there for our benefit, but the benefit is for everybody.”
Topics:data-and-communication,computer-science,technological-artificial-intelligence,information-technology,sciences,federal—state-issues,information,technology,data,research,science-and.technology,melbourne-3000More stories from New South Wales