mirror of
https://github.com/huggingface/transformers.git
synced 2025-07-31 02:02:21 +06:00
![]() * First draft * More improvements * Improve conversion script * Fix init weights for layer norm * Fix correct model for conversion script * Don't tie input and output embeddings * Add print statements for debugging * Add print statements for debugging * Fix vocab size of model * Improve documentation, remove fast tokenizer * Add ImageGPTForImageClassification, improve docs * Fix docs issue * Set verbosity level back to info * Improve tests * Fix tests and add figure * Delete tokenizer file * Remove ImageGPTTokenizer from init files * Remove ImageGPTLayer from init files * Remove ImageGPT tokenizer from docs * First draft of ImageGPTFeatureExtractor * Fix typo * Fix bug * More improvements * Apply suggestions from code review, add tests for feature extractor * Fix layernorm * Update save_pretrained method * Fix issue * Make all tests of ImageGPTFeatureExtractor pass * Update code examples * Rename model inputs to pixel_values * Improve code examples * Update init_weights to post_init * Fix post_init |
||
---|---|---|
.. | ||
course_banner.png | ||
ImageGPT.png | ||
local_attention_mask.png | ||
parallelism-deepspeed-3d.png | ||
parallelism-flexflow.jpeg | ||
parallelism-gpipe-bubble.png | ||
parallelism-sagemaker-interleaved-pipeline.png | ||
parallelism-tp-independent-gelu.png | ||
parallelism-tp-parallel_gemm.png | ||
parallelism-tp-parallel_self_attention.png | ||
parallelism-tp-parallel_shard_processing.png | ||
parallelism-zero-dp-pp.png | ||
parallelism-zero.png | ||
ppl_chunked.gif | ||
ppl_full.gif | ||
ppl_sliding.gif | ||
transformers_logo_name.png | ||
transformers_overview.png | ||
warmup_constant_schedule.png | ||
warmup_cosine_hard_restarts_schedule.png | ||
warmup_cosine_schedule.png | ||
warmup_cosine_warm_restarts_schedule.png | ||
warmup_linear_schedule.png |