Giter VIP home page Giter VIP logo

IST Austria Distributed Algorithms and Systems Lab's Projects

acdc icon acdc

Code for reproducing "AC/DC: Alternating Compressed/DeCompressed Training of Deep Neural Networks" (NeurIPS 2021)

cap icon cap

Repository for Correlation Aware Prune (NeurIPS23) source and experimental code

cram icon cram

Code for reproducing the results from "CrAM: A Compression-Aware Minimizer" accepted at ICLR 2023

distiller icon distiller

Neural Network Distiller by Intel AI Lab: a Python package for neural network compression research. https://nervanasystems.github.io/distiller

efcp icon efcp

The repository contains code to reproduce the experiments from our paper Error Feedback Can Accurately Compress Preconditioners available below:

gptq icon gptq

Code for the ICLR 2023 paper "GPTQ: Accurate Post-training Quantization of Generative Pretrained Transformers".

horovod icon horovod

Distributed training framework for TensorFlow, Keras, PyTorch, and Apache MXNet.

kdvr icon kdvr

Code for the experiments in Knowledge Distillation Performs Partial Variance Reduction, NeurIPS 2023

m-fac icon m-fac

Efficient reference implementations of the static & dynamic M-FAC algorithms (for pruning and optimization)

marlin icon marlin

FP16xINT4 LLM inference kernel that can achieve near-ideal ~4x speedups up to medium batchsizes of 16-32 tokens.

obc icon obc

Code for the NeurIPS 2022 paper "Optimal Brain Compression: A Framework for Accurate Post-Training Quantization and Pruning".

peft-rosa icon peft-rosa

A fork of the PEFT library, supporting Robust Adaptation (RoSA)

qigen icon qigen

Repository for CPU Kernel Generation for LLM Inference

qmoe icon qmoe

Code for the paper "QMoE: Practical Sub-1-Bit Compression of Trillion-Parameter Models".

qrgd icon qrgd

Repository for the implementation of "Distributed Principal Component Analysis with Limited Communication" (Alimisis et al., NeurIPS 2021). Parts of this code were originally based on code from "Communication-Efficient Distributed PCA by Riemannian Optimization" (Huang and Pan, ICML 2020).

quik icon quik

Repository for the QUIK project, enabling the use of 4bit kernels for generative inference

smart-quantizer icon smart-quantizer

Repository for Vitaly's implementation of the distribution-adaptive quantizer

sparse-imagenet-transfer icon sparse-imagenet-transfer

Code for reproducing the results in "How Well do Sparse Imagenet Models Transfer?", presented at CVPR 2022

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.