Friday, February 7, 2025

Artificial Intelligence news

These documents are influencing...

Reports from the US Government Accountability Office on improper federal payments in...

Reframing digital transformation through...

Enterprise adoption of generative AI technologies has undergone explosive growth in the...

An AI chatbot told...

For the past five months, Al Nowatzki has been talking to an...

What’s next for smart...

MIT Technology Review’s What’s Next series looks across industries, trends, and technologies...
HomeMachine LearningPDP: Parameter-free Differentiable...

PDP: Parameter-free Differentiable Pruning is All You Need



DNN pruning is a popular way to reduce the size of a model, improve the inference latency, and minimize the power consumption on DNN accelerators. However, existing approaches might be too complex, expensive or ineffective to apply to a variety of vision/language tasks, DNN architectures and to honor structured pruning constraints. In this paper, we propose an efficient yet effective train-time pruning scheme, Parameter-free Differentiable Pruning (PDP), which offers state-of-the-art qualities in model size, accuracy, and training cost. PDP uses a dynamic function of weights during training to…



Article Source link and Credit

Continue reading

Mitigating Hallucinated Translations in Large Language Models with Hallucination-focused Preference Optimization

Machine Translation (MT) is undergoing a paradigm shift, with systems based on fine-tuned large language models (LLM) becoming increasingly competitive with traditional encoder-decoder models trained specifically for translation tasks. However, LLM-based systems are at a higher risk of...

Parameters vs FLOPs: Scaling Laws for Optimal Sparsity for Mixture-of-Experts Language Models

Scaling the capacity of language models has consistently proven to be a reliable approach for improving performance and unlocking new capabilities. Capacity can be primarily defined by two dimensions: the number of model parameters and the compute per example. While...

Compact Neural TTS Voices for Accessibility

Contemporary text-to-speech solutions for accessibility applications can typically be classified into two categories: (i) device-based statistical parametric speech synthesis (SPSS) or unit selection (USEL) and (ii) cloud-based neural TTS. SPSS and USEL offer low latency and low disk...