Buckets:
| # Optimum.Neuron | |
| ## Docs | |
| - [Optimum Neuron Container](https://huggingface.co/docs/optimum.neuron/v0.4.3/containers.md) | |
| - [Supported architectures](https://huggingface.co/docs/optimum.neuron/v0.4.3/supported_architectures.md) | |
| - [๐ค Optimum Neuron](https://huggingface.co/docs/optimum.neuron/v0.4.3/index.md) | |
| - [EC2 Setup](https://huggingface.co/docs/optimum.neuron/v0.4.3/ec2-setup.md) | |
| - [Quickstart](https://huggingface.co/docs/optimum.neuron/v0.4.3/quickstart.md) | |
| - [Getting started with AWS Trainium and Hugging Face Transformers](https://huggingface.co/docs/optimum.neuron/v0.4.3/training_tutorials/fine_tune_bert.md) | |
| - [๐ Tutorials: How To Fine-tune & Run LLMs](https://huggingface.co/docs/optimum.neuron/v0.4.3/training_tutorials/finetune_llms_overview.md) | |
| - [๐ Continuous Pretraining of Llama 3.2 1B on SageMaker Hyperpod with Pre-built Containers](https://huggingface.co/docs/optimum.neuron/v0.4.3/training_tutorials/pretraining_hyperpod_llm.md) | |
| - [๐ Fine-Tune Qwen3 8B with LoRA](https://huggingface.co/docs/optimum.neuron/v0.4.3/training_tutorials/finetune_qwen3.md) | |
| - [๐ Instruction Fine-Tuning of Llama 3.1 8B with LoRA](https://huggingface.co/docs/optimum.neuron/v0.4.3/training_tutorials/finetune_llama.md) | |
| - [Llama-3.3-70b performance on AWS Inferentia2 (Latency & Throughput)](https://huggingface.co/docs/optimum.neuron/v0.4.3/benchmarks/inferentia-llama3.3-70b.md) | |
| - [Llama-3.1-8b performance on AWS Inferentia2 (Latency & Throughput)](https://huggingface.co/docs/optimum.neuron/v0.4.3/benchmarks/inferentia-llama3.1-8b.md) | |
| - [Introduction](https://huggingface.co/docs/optimum.neuron/v0.4.3/guides/benchmark.md) | |
| - [Inference pipelines with AWS Neuron (Inf2/Trn1)](https://huggingface.co/docs/optimum.neuron/v0.4.3/guides/pipelines.md) | |
| - [optimum-neuron plugin for vLLM](https://huggingface.co/docs/optimum.neuron/v0.4.3/guides/vllm_plugin.md) | |
| - [NeuronX Text-generation-inference for AWS inferentia2](https://huggingface.co/docs/optimum.neuron/v0.4.3/guides/neuronx_tgi.md) | |
| - [Neuron Model Cache](https://huggingface.co/docs/optimum.neuron/v0.4.3/guides/cache_system.md) | |
| - [Distributed Training with `optimum-neuron`](https://huggingface.co/docs/optimum.neuron/v0.4.3/guides/distributed_training.md) | |
| - [Export a model to Neuron](https://huggingface.co/docs/optimum.neuron/v0.4.3/guides/export_model.md) | |
| - [Models](https://huggingface.co/docs/optimum.neuron/v0.4.3/model_doc/modeling_auto.md) | |
| - [Sentence Transformers ๐ค](https://huggingface.co/docs/optimum.neuron/v0.4.3/model_doc/sentence_transformers/overview.md) | |
| - [Latent Consistency Models](https://huggingface.co/docs/optimum.neuron/v0.4.3/model_doc/diffusers/lcm.md) | |
| - [Stable Diffusion XL](https://huggingface.co/docs/optimum.neuron/v0.4.3/model_doc/diffusers/stable_diffusion_xl.md) | |
| - [Load adapters](https://huggingface.co/docs/optimum.neuron/v0.4.3/model_doc/diffusers/lora.md) | |
| - [IP-Adapter](https://huggingface.co/docs/optimum.neuron/v0.4.3/model_doc/diffusers/ip_adapter.md) | |
| - [Flux](https://huggingface.co/docs/optimum.neuron/v0.4.3/model_doc/diffusers/flux.md) | |
| - [Stable Diffusion](https://huggingface.co/docs/optimum.neuron/v0.4.3/model_doc/diffusers/stable_diffusion.md) | |
| - [Stable Diffusion XL Turbo](https://huggingface.co/docs/optimum.neuron/v0.4.3/model_doc/diffusers/sdxl_turbo.md) | |
| - [ControlNet](https://huggingface.co/docs/optimum.neuron/v0.4.3/model_doc/diffusers/controlnet.md) | |
| - [PixArt-ฮฑ](https://huggingface.co/docs/optimum.neuron/v0.4.3/model_doc/diffusers/pixart_alpha.md) | |
| - [PixArt-ฮฃ](https://huggingface.co/docs/optimum.neuron/v0.4.3/model_doc/diffusers/pixart_sigma.md) | |
| - [InstructPix2Pix](https://huggingface.co/docs/optimum.neuron/v0.4.3/model_doc/diffusers/pix2pix.md) | |
| - [YOLOS](https://huggingface.co/docs/optimum.neuron/v0.4.3/model_doc/transformers/yolos.md) | |
| - [CLIP](https://huggingface.co/docs/optimum.neuron/v0.4.3/model_doc/transformers/clip.md) | |
| - [Whisper](https://huggingface.co/docs/optimum.neuron/v0.4.3/model_doc/transformers/whisper.md) | |
| - [BERT](https://huggingface.co/docs/optimum.neuron/v0.4.3/model_doc/transformers/bert.md) | |
| - [Setting up your development environment](https://huggingface.co/docs/optimum.neuron/v0.4.3/contribute/dev_environment.md) | |
| - [Contributing Custom Models for Training](https://huggingface.co/docs/optimum.neuron/v0.4.3/contribute/contribute_for_training.md) | |
| - [Adding support for new architectures](https://huggingface.co/docs/optimum.neuron/v0.4.3/contribute/contribute_for_inference.md) | |
| - [Create your own chatbot with llama-2-13B on AWS Inferentia](https://huggingface.co/docs/optimum.neuron/v0.4.3/inference_tutorials/llama2-13b-chatbot.md) | |
| - [Deploy Llama 3.3 70B on AWS Inferentia2](https://huggingface.co/docs/optimum.neuron/v0.4.3/inference_tutorials/deploy-llama-3-3-70b.md) | |
| - [Sentence Transformers on AWS Inferentia with Optimum Neuron](https://huggingface.co/docs/optimum.neuron/v0.4.3/inference_tutorials/sentence_transformers.md) | |
| - [Deploy Mixtral 8x7B on AWS Inferentia2](https://huggingface.co/docs/optimum.neuron/v0.4.3/inference_tutorials/deploy-mixtral-8x7b.md) | |
| - [Notebooks](https://huggingface.co/docs/optimum.neuron/v0.4.3/inference_tutorials/notebooks.md) | |
| - [NeuronTrainer](https://huggingface.co/docs/optimum.neuron/v0.4.3/training_api/trainer.md) | |
| - [LoRA for Neuron](https://huggingface.co/docs/optimum.neuron/v0.4.3/training_api/lora.md) | |
| - [Neuron TRL Trainers](https://huggingface.co/docs/optimum.neuron/v0.4.3/training_api/trl_trainers.md) | |
| - [Model Weight Transformation Specs](https://huggingface.co/docs/optimum.neuron/v0.4.3/training_api/transformations.md) | |
Xet Storage Details
- Size:
- 5.55 kB
- Xet hash:
- 764b7376bc992780a0aed06a85e326f4c646ac196d6cc87ebdbf17f53cedfb6a
ยท
Xet efficiently stores files, intelligently splitting them into unique chunks and accelerating uploads and downloads. More info.