Buckets:

rtrm's picture
download
raw
5.55 kB
# Optimum.Neuron
## Docs
- [Optimum Neuron Container](https://huggingface.co/docs/optimum.neuron/v0.4.3/containers.md)
- [Supported architectures](https://huggingface.co/docs/optimum.neuron/v0.4.3/supported_architectures.md)
- [๐Ÿค— Optimum Neuron](https://huggingface.co/docs/optimum.neuron/v0.4.3/index.md)
- [EC2 Setup](https://huggingface.co/docs/optimum.neuron/v0.4.3/ec2-setup.md)
- [Quickstart](https://huggingface.co/docs/optimum.neuron/v0.4.3/quickstart.md)
- [Getting started with AWS Trainium and Hugging Face Transformers](https://huggingface.co/docs/optimum.neuron/v0.4.3/training_tutorials/fine_tune_bert.md)
- [๐Ÿš€ Tutorials: How To Fine-tune & Run LLMs](https://huggingface.co/docs/optimum.neuron/v0.4.3/training_tutorials/finetune_llms_overview.md)
- [๐Ÿš€ Continuous Pretraining of Llama 3.2 1B on SageMaker Hyperpod with Pre-built Containers](https://huggingface.co/docs/optimum.neuron/v0.4.3/training_tutorials/pretraining_hyperpod_llm.md)
- [๐Ÿš€ Fine-Tune Qwen3 8B with LoRA](https://huggingface.co/docs/optimum.neuron/v0.4.3/training_tutorials/finetune_qwen3.md)
- [๐Ÿš€ Instruction Fine-Tuning of Llama 3.1 8B with LoRA](https://huggingface.co/docs/optimum.neuron/v0.4.3/training_tutorials/finetune_llama.md)
- [Llama-3.3-70b performance on AWS Inferentia2 (Latency & Throughput)](https://huggingface.co/docs/optimum.neuron/v0.4.3/benchmarks/inferentia-llama3.3-70b.md)
- [Llama-3.1-8b performance on AWS Inferentia2 (Latency & Throughput)](https://huggingface.co/docs/optimum.neuron/v0.4.3/benchmarks/inferentia-llama3.1-8b.md)
- [Introduction](https://huggingface.co/docs/optimum.neuron/v0.4.3/guides/benchmark.md)
- [Inference pipelines with AWS Neuron (Inf2/Trn1)](https://huggingface.co/docs/optimum.neuron/v0.4.3/guides/pipelines.md)
- [optimum-neuron plugin for vLLM](https://huggingface.co/docs/optimum.neuron/v0.4.3/guides/vllm_plugin.md)
- [NeuronX Text-generation-inference for AWS inferentia2](https://huggingface.co/docs/optimum.neuron/v0.4.3/guides/neuronx_tgi.md)
- [Neuron Model Cache](https://huggingface.co/docs/optimum.neuron/v0.4.3/guides/cache_system.md)
- [Distributed Training with `optimum-neuron`](https://huggingface.co/docs/optimum.neuron/v0.4.3/guides/distributed_training.md)
- [Export a model to Neuron](https://huggingface.co/docs/optimum.neuron/v0.4.3/guides/export_model.md)
- [Models](https://huggingface.co/docs/optimum.neuron/v0.4.3/model_doc/modeling_auto.md)
- [Sentence Transformers ๐Ÿค—](https://huggingface.co/docs/optimum.neuron/v0.4.3/model_doc/sentence_transformers/overview.md)
- [Latent Consistency Models](https://huggingface.co/docs/optimum.neuron/v0.4.3/model_doc/diffusers/lcm.md)
- [Stable Diffusion XL](https://huggingface.co/docs/optimum.neuron/v0.4.3/model_doc/diffusers/stable_diffusion_xl.md)
- [Load adapters](https://huggingface.co/docs/optimum.neuron/v0.4.3/model_doc/diffusers/lora.md)
- [IP-Adapter](https://huggingface.co/docs/optimum.neuron/v0.4.3/model_doc/diffusers/ip_adapter.md)
- [Flux](https://huggingface.co/docs/optimum.neuron/v0.4.3/model_doc/diffusers/flux.md)
- [Stable Diffusion](https://huggingface.co/docs/optimum.neuron/v0.4.3/model_doc/diffusers/stable_diffusion.md)
- [Stable Diffusion XL Turbo](https://huggingface.co/docs/optimum.neuron/v0.4.3/model_doc/diffusers/sdxl_turbo.md)
- [ControlNet](https://huggingface.co/docs/optimum.neuron/v0.4.3/model_doc/diffusers/controlnet.md)
- [PixArt-ฮฑ](https://huggingface.co/docs/optimum.neuron/v0.4.3/model_doc/diffusers/pixart_alpha.md)
- [PixArt-ฮฃ](https://huggingface.co/docs/optimum.neuron/v0.4.3/model_doc/diffusers/pixart_sigma.md)
- [InstructPix2Pix](https://huggingface.co/docs/optimum.neuron/v0.4.3/model_doc/diffusers/pix2pix.md)
- [YOLOS](https://huggingface.co/docs/optimum.neuron/v0.4.3/model_doc/transformers/yolos.md)
- [CLIP](https://huggingface.co/docs/optimum.neuron/v0.4.3/model_doc/transformers/clip.md)
- [Whisper](https://huggingface.co/docs/optimum.neuron/v0.4.3/model_doc/transformers/whisper.md)
- [BERT](https://huggingface.co/docs/optimum.neuron/v0.4.3/model_doc/transformers/bert.md)
- [Setting up your development environment](https://huggingface.co/docs/optimum.neuron/v0.4.3/contribute/dev_environment.md)
- [Contributing Custom Models for Training](https://huggingface.co/docs/optimum.neuron/v0.4.3/contribute/contribute_for_training.md)
- [Adding support for new architectures](https://huggingface.co/docs/optimum.neuron/v0.4.3/contribute/contribute_for_inference.md)
- [Create your own chatbot with llama-2-13B on AWS Inferentia](https://huggingface.co/docs/optimum.neuron/v0.4.3/inference_tutorials/llama2-13b-chatbot.md)
- [Deploy Llama 3.3 70B on AWS Inferentia2](https://huggingface.co/docs/optimum.neuron/v0.4.3/inference_tutorials/deploy-llama-3-3-70b.md)
- [Sentence Transformers on AWS Inferentia with Optimum Neuron](https://huggingface.co/docs/optimum.neuron/v0.4.3/inference_tutorials/sentence_transformers.md)
- [Deploy Mixtral 8x7B on AWS Inferentia2](https://huggingface.co/docs/optimum.neuron/v0.4.3/inference_tutorials/deploy-mixtral-8x7b.md)
- [Notebooks](https://huggingface.co/docs/optimum.neuron/v0.4.3/inference_tutorials/notebooks.md)
- [NeuronTrainer](https://huggingface.co/docs/optimum.neuron/v0.4.3/training_api/trainer.md)
- [LoRA for Neuron](https://huggingface.co/docs/optimum.neuron/v0.4.3/training_api/lora.md)
- [Neuron TRL Trainers](https://huggingface.co/docs/optimum.neuron/v0.4.3/training_api/trl_trainers.md)
- [Model Weight Transformation Specs](https://huggingface.co/docs/optimum.neuron/v0.4.3/training_api/transformations.md)

Xet Storage Details

Size:
5.55 kB
ยท
Xet hash:
764b7376bc992780a0aed06a85e326f4c646ac196d6cc87ebdbf17f53cedfb6a

Xet efficiently stores files, intelligently splitting them into unique chunks and accelerating uploads and downloads. More info.