这是indexloc提供的服务,不要输入任何密码
Skip to main content

Showing 1–21 of 21 results for author: Yvinec, E

Searching in archive cs. Search in all archives.
.
  1. arXiv:2507.06261  [pdf, ps, other

    cs.CL cs.AI

    Gemini 2.5: Pushing the Frontier with Advanced Reasoning, Multimodality, Long Context, and Next Generation Agentic Capabilities

    Authors: Gheorghe Comanici, Eric Bieber, Mike Schaekermann, Ice Pasupat, Noveen Sachdeva, Inderjit Dhillon, Marcel Blistein, Ori Ram, Dan Zhang, Evan Rosen, Luke Marris, Sam Petulla, Colin Gaffney, Asaf Aharoni, Nathan Lintz, Tiago Cardal Pais, Henrik Jacobsson, Idan Szpektor, Nan-Jiang Jiang, Krishna Haridasan, Ahmed Omran, Nikunj Saunshi, Dara Bahri, Gaurav Mishra, Eric Chu , et al. (3284 additional authors not shown)

    Abstract: In this report, we introduce the Gemini 2.X model family: Gemini 2.5 Pro and Gemini 2.5 Flash, as well as our earlier Gemini 2.0 Flash and Flash-Lite models. Gemini 2.5 Pro is our most capable model yet, achieving SoTA performance on frontier coding and reasoning benchmarks. In addition to its incredible coding and reasoning skills, Gemini 2.5 Pro is a thinking model that excels at multimodal unde… ▽ More

    Submitted 22 July, 2025; v1 submitted 7 July, 2025; originally announced July 2025.

    Comments: 72 pages, 17 figures

  2. arXiv:2507.05201  [pdf, ps, other

    cs.AI cs.CL cs.CV

    MedGemma Technical Report

    Authors: Andrew Sellergren, Sahar Kazemzadeh, Tiam Jaroensri, Atilla Kiraly, Madeleine Traverse, Timo Kohlberger, Shawn Xu, Fayaz Jamil, Cían Hughes, Charles Lau, Justin Chen, Fereshteh Mahvar, Liron Yatziv, Tiffany Chen, Bram Sterling, Stefanie Anna Baby, Susanna Maria Baby, Jeremy Lai, Samuel Schmidgall, Lu Yang, Kejia Chen, Per Bjornsson, Shashir Reddy, Ryan Brush, Kenneth Philbrick , et al. (56 additional authors not shown)

    Abstract: Artificial intelligence (AI) has significant potential in healthcare applications, but its training and deployment faces challenges due to healthcare's diverse data, complex tasks, and the need to preserve privacy. Foundation models that perform well on medical tasks and require less task-specific tuning data are critical to accelerate the development of healthcare AI applications. We introduce Me… ▽ More

    Submitted 12 July, 2025; v1 submitted 7 July, 2025; originally announced July 2025.

  3. arXiv:2503.19786  [pdf, other

    cs.CL cs.AI

    Gemma 3 Technical Report

    Authors: Gemma Team, Aishwarya Kamath, Johan Ferret, Shreya Pathak, Nino Vieillard, Ramona Merhej, Sarah Perrin, Tatiana Matejovicova, Alexandre Ramé, Morgane Rivière, Louis Rouillard, Thomas Mesnard, Geoffrey Cideron, Jean-bastien Grill, Sabela Ramos, Edouard Yvinec, Michelle Casbon, Etienne Pot, Ivo Penchev, Gaël Liu, Francesco Visin, Kathleen Kenealy, Lucas Beyer, Xiaohai Zhai, Anton Tsitsulin , et al. (191 additional authors not shown)

    Abstract: We introduce Gemma 3, a multimodal addition to the Gemma family of lightweight open models, ranging in scale from 1 to 27 billion parameters. This version introduces vision understanding abilities, a wider coverage of languages and longer context - at least 128K tokens. We also change the architecture of the model to reduce the KV-cache memory that tends to explode with long context. This is achie… ▽ More

    Submitted 25 March, 2025; originally announced March 2025.

  4. arXiv:2405.20505  [pdf, other

    cs.CL cs.LG

    SPOT: Text Source Prediction from Originality Score Thresholding

    Authors: Edouard Yvinec, Gabriel Kasser

    Abstract: The wide acceptance of large language models (LLMs) has unlocked new applications and social risks. Popular countermeasures aim at detecting misinformation, usually involve domain specific models trained to recognize the relevance of any information. Instead of evaluating the validity of the information, we propose to investigate LLM generated text from the perspective of trust. In this study, we… ▽ More

    Submitted 30 May, 2024; originally announced May 2024.

  5. arXiv:2311.15806  [pdf, other

    cs.CV

    PIPE : Parallelized Inference Through Post-Training Quantization Ensembling of Residual Expansions

    Authors: Edouard Yvinec, Arnaud Dapogny, Kevin Bailly

    Abstract: Deep neural networks (DNNs) are ubiquitous in computer vision and natural language processing, but suffer from high inference cost. This problem can be addressed by quantization, which consists in converting floating point perations into a lower bit-width format. With the growing concerns on privacy rights, we focus our efforts on data-free methods. However, such techniques suffer from their lack… ▽ More

    Submitted 27 November, 2023; originally announced November 2023.

    Comments: arXiv admin note: substantial text overlap with arXiv:2203.14645

  6. arXiv:2311.10549  [pdf, other

    cs.CV

    Archtree: on-the-fly tree-structured exploration for latency-aware pruning of deep neural networks

    Authors: Rémi Ouazan Reboul, Edouard Yvinec, Arnaud Dapogny, Kevin Bailly

    Abstract: Deep neural networks (DNNs) have become ubiquitous in addressing a number of problems, particularly in computer vision. However, DNN inference is computationally intensive, which can be prohibitive e.g. when considering edge devices. To solve this problem, a popular solution is DNN pruning, and more so structured pruning, where coherent computational blocks (e.g. channels for convolutional network… ▽ More

    Submitted 17 November, 2023; originally announced November 2023.

    Comments: 10 pages, 7 figures

  7. arXiv:2309.17361  [pdf, other

    cs.CV

    Network Memory Footprint Compression Through Jointly Learnable Codebooks and Mappings

    Authors: Edouard Yvinec, Arnaud Dapogny, Kevin Bailly

    Abstract: The massive interest in deep neural networks (DNNs) for both computer vision and natural language processing has been sparked by the growth in computational power. However, this led to an increase in the memory footprint, to a point where it can be challenging to simply load a model on commodity devices such as mobile phones. To address this limitation, quantization is a favored solution as it map… ▽ More

    Submitted 29 September, 2023; originally announced September 2023.

  8. arXiv:2308.07662  [pdf, other

    cs.LG cs.CV

    Gradient-Based Post-Training Quantization: Challenging the Status Quo

    Authors: Edouard Yvinec, Arnaud Dapogny, Kevin Bailly

    Abstract: Quantization has become a crucial step for the efficient deployment of deep neural networks, where floating point operations are converted to simpler fixed point operations. In its most naive form, it simply consists in a combination of scaling and rounding transformations, leading to either a limited compression rate or a significant accuracy drop. Recently, Gradient-based post-training quantizat… ▽ More

    Submitted 15 August, 2023; originally announced August 2023.

  9. arXiv:2308.05600  [pdf, other

    cs.LG cs.CV

    NUPES : Non-Uniform Post-Training Quantization via Power Exponent Search

    Authors: Edouard Yvinec, Arnaud Dapogny, Kevin Bailly

    Abstract: Deep neural network (DNN) deployment has been confined to larger hardware devices due to their expensive computational requirements. This challenge has recently reached another scale with the emergence of large language models (LLMs). In order to reduce both their memory footprint and latency, a promising technique is quantization. It consists in converting floating point representations to low bi… ▽ More

    Submitted 10 August, 2023; originally announced August 2023.

  10. arXiv:2308.04753  [pdf, other

    cs.CV

    SAfER: Layer-Level Sensitivity Assessment for Efficient and Robust Neural Network Inference

    Authors: Edouard Yvinec, Arnaud Dapogny, Kevin Bailly, Xavier Fischer

    Abstract: Deep neural networks (DNNs) demonstrate outstanding performance across most computer vision tasks. Some critical applications, such as autonomous driving or medical imaging, also require investigation into their behavior and the reasons behind the decisions they make. In this vein, DNN attribution consists in studying the relationship between the predictions of a DNN and its inputs. Attribution me… ▽ More

    Submitted 8 December, 2023; v1 submitted 9 August, 2023; originally announced August 2023.

  11. arXiv:2306.17442  [pdf, other

    cs.CV

    Designing strong baselines for ternary neural network quantization through support and mass equalization

    Authors: Edouard Yvinec, Arnaud Dapogny, Kevin Bailly

    Abstract: Deep neural networks (DNNs) offer the highest performance in a wide range of applications in computer vision. These results rely on over-parameterized backbones, which are expensive to run. This computational burden can be dramatically reduced by quantizing (in either data-free (DFQ), post-training (PTQ) or quantization-aware training (QAT) scenarios) floating point values to ternary values (2 bit… ▽ More

    Submitted 30 June, 2023; originally announced June 2023.

    Journal ref: ICIP 2023

  12. arXiv:2303.11803  [pdf, other

    cs.CV

    Fighting over-fitting with quantization for learning deep neural networks on noisy labels

    Authors: Gauthier Tallec, Edouard Yvinec, Arnaud Dapogny, Kevin Bailly

    Abstract: The rising performance of deep neural networks is often empirically attributed to an increase in the available computational power, which allows complex models to be trained upon large amounts of annotated data. However, increased model complexity leads to costly deployment of modern neural networks, while gathering such amounts of data requires huge costs to avoid label noise. In this work, we st… ▽ More

    Submitted 21 March, 2023; originally announced March 2023.

  13. arXiv:2301.09858  [pdf, other

    cs.CV

    PowerQuant: Automorphism Search for Non-Uniform Quantization

    Authors: Edouard Yvinec, Arnaud Dapogny, Matthieu Cord, Kevin Bailly

    Abstract: Deep neural networks (DNNs) are nowadays ubiquitous in many domains such as computer vision. However, due to their high latency, the deployment of DNNs hinges on the development of compression techniques such as quantization which consists in lowering the number of bits used to encode the weights and activations. Growing concerns for privacy and security have motivated the development of data-free… ▽ More

    Submitted 24 January, 2023; originally announced January 2023.

  14. arXiv:2207.04089  [pdf, other

    cs.CV

    SInGE: Sparsity via Integrated Gradients Estimation of Neuron Relevance

    Authors: Edouard Yvinec, Arnaud Dapogny, Matthieu Cord, Kevin Bailly

    Abstract: The leap in performance in state-of-the-art computer vision methods is attributed to the development of deep neural networks. However it often comes at a computational price which may hinder their deployment. To alleviate this limitation, structured pruning is a well known technique which consists in removing channels, neurons or filters, and is commonly applied in order to produce more compact mo… ▽ More

    Submitted 8 July, 2022; originally announced July 2022.

  15. arXiv:2203.14646  [pdf, other

    cs.LG cs.CV

    To Fold or Not to Fold: a Necessary and Sufficient Condition on Batch-Normalization Layers Folding

    Authors: Edouard Yvinec, Arnaud Dapogny, Kevin Bailly

    Abstract: Batch-Normalization (BN) layers have become fundamental components in the evermore complex deep neural network architectures. Such models require acceleration processes for deployment on edge devices. However, BN layers add computation bottlenecks due to the sequential operation processing: thus, a key, yet often overlooked component of the acceleration process is BN layers folding. In this paper,… ▽ More

    Submitted 28 March, 2022; originally announced March 2022.

  16. arXiv:2203.14645  [pdf, other

    cs.CV

    REx: Data-Free Residual Quantization Error Expansion

    Authors: Edouard Yvinec, Arnaud Dapgony, Matthieu Cord, Kevin Bailly

    Abstract: Deep neural networks (DNNs) are ubiquitous in computer vision and natural language processing, but suffer from high inference cost. This problem can be addressed by quantization, which consists in converting floating point operations into a lower bit-width format. With the growing concerns on privacy rights, we focus our efforts on data-free methods. However, such techniques suffer from their lack… ▽ More

    Submitted 29 May, 2023; v1 submitted 28 March, 2022; originally announced March 2022.

  17. arXiv:2203.14642  [pdf, other

    cs.CV

    SPIQ: Data-Free Per-Channel Static Input Quantization

    Authors: Edouard Yvinec, Arnaud Dapogny, Matthieu Cord, Kevin Bailly

    Abstract: Computationally expensive neural networks are ubiquitous in computer vision and solutions for efficient inference have drawn a growing attention in the machine learning community. Examples of such solutions comprise quantization, i.e. converting the processing values (weights and inputs) from floating point into integers e.g. int8 or int4. Concurrently, the rise of privacy concerns motivated the s… ▽ More

    Submitted 28 March, 2022; originally announced March 2022.

  18. arXiv:2203.12531  [pdf, ps, other

    cs.CV

    Multi-label Transformer for Action Unit Detection

    Authors: Gauthier Tallec, Edouard Yvinec, Arnaud Dapogny, Kevin Bailly

    Abstract: Action Unit (AU) Detection is the branch of affective computing that aims at recognizing unitary facial muscular movements. It is key to unlock unbiased computational face representations and has therefore aroused great interest in the past few years. One of the main obstacles toward building efficient deep learning based AU detection system is the lack of wide facial image databases annotated by… ▽ More

    Submitted 12 December, 2022; v1 submitted 23 March, 2022; originally announced March 2022.

  19. arXiv:2110.01397  [pdf, other

    cs.LG cs.CV

    RED++ : Data-Free Pruning of Deep Neural Networks via Input Splitting and Output Merging

    Authors: Edouard Yvinec, Arnaud Dapogny, Matthieu Cord, Kevin Bailly

    Abstract: Pruning Deep Neural Networks (DNNs) is a prominent field of study in the goal of inference runtime acceleration. In this paper, we introduce a novel data-free pruning protocol RED++. Only requiring a trained neural network, and not specific to DNN architecture, we exploit an adaptive data-free scalar hashing which exhibits redundancies among neuron weight values. We study the theoretical and empir… ▽ More

    Submitted 30 September, 2021; originally announced October 2021.

    Comments: 18 pages, 10 figures

  20. arXiv:2105.14797  [pdf, other

    cs.CV eess.IV

    RED : Looking for Redundancies for Data-Free Structured Compression of Deep Neural Networks

    Authors: Edouard Yvinec, Arnaud Dapogny, Matthieu Cord, Kevin Bailly

    Abstract: Deep Neural Networks (DNNs) are ubiquitous in today's computer vision land-scape, despite involving considerable computational costs. The mainstream approaches for runtime acceleration consist in pruning connections (unstructured pruning) or, better, filters (structured pruning), both often requiring data to re-train the model. In this paper, we present RED, a data-free structured, unified approac… ▽ More

    Submitted 31 May, 2021; originally announced May 2021.

  21. arXiv:2004.07098  [pdf, other

    cs.CV

    DeeSCo: Deep heterogeneous ensemble with Stochastic Combinatory loss for gaze estimation

    Authors: Edouard Yvinec, Arnaud Dapogny, Kévin Bailly

    Abstract: From medical research to gaming applications, gaze estimation is becoming a valuable tool. While there exists a number of hardware-based solutions, recent deep learning-based approaches, coupled with the availability of large-scale databases, have allowed to provide a precise gaze estimate using only consumer sensors. However, there remains a number of questions, regarding the problem formulation,… ▽ More

    Submitted 15 April, 2020; originally announced April 2020.

    Comments: 7 pages, 6 figures, FG 2020