forked from erfanzar/FJFormer
-
Notifications
You must be signed in to change notification settings - Fork 0
/
mkdocs.yml
78 lines (75 loc) · 2.75 KB
/
mkdocs.yml
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
nav:
- Bits:
- Bits: generated-bits-bits.md
- Calibration: generated-bits-calibration.md
- Config: generated-bits-config.md
- Int Numerics: generated-bits-int_numerics.md
- No Numerics: generated-bits-no_numerics.md
- Numerics: generated-bits-numerics.md
- Q Dot General: generated-bits-q_dot_general.md
- Q Flax: generated-bits-q_flax.md
- Stochastic Rounding: generated-bits-stochastic_rounding.md
- Checkpoint:
- Load: generated-checkpoint-_load.md
- Streamer: generated-checkpoint-streamer.md
- Datasets:
- Datasets: generated-datasets-datasets.md
- Func:
- Func: generated-func-_func.md
- Loss Func: generated-func-loss_func.md
- Home: index.md
- Monitor:
- Tracker: generated-monitor-tracker.md
- Optimizers:
- Adafactor: generated-optimizers-adafactor.md
- Adamw: generated-optimizers-adamw.md
- Lion: generated-optimizers-lion.md
- Optimizer Utils: generated-optimizers-optimizer_utils.md
- Rmsprop: generated-optimizers-rmsprop.md
- Pallas Operations:
- Efficient Attention:
- Efficient Attention: generated-pallas_operations-efficient_attention-efficient_attention.md
- Flash Attention:
- Gpu:
- Jax Flash Attn Gpu: generated-pallas_operations-flash_attention-gpu-jax_flash_attn_gpu.md
- Tpu:
- Jax Flash Attn Tpu: generated-pallas_operations-flash_attention-tpu-jax_flash_attn_tpu.md
- Layer Norm:
- Gpu:
- Layer Norm: generated-pallas_operations-layer_norm-gpu-layer_norm.md
- Rms Norm:
- Gpu:
- Rms Norm: generated-pallas_operations-rms_norm-gpu-rms_norm.md
- Softmax:
- Gpu:
- Softmax: generated-pallas_operations-softmax-gpu-softmax.md
- Splash Attention:
- Tpu:
- Splash Attention Kernel: generated-pallas_operations-splash_attention-tpu-splash_attention_kernel.md
- Splash Attention Mask: generated-pallas_operations-splash_attention-tpu-splash_attention_mask.md
- Splash Attention Mask Info: generated-pallas_operations-splash_attention-tpu-splash_attention_mask_info.md
- Partition Utils:
- Mesh Utils: generated-partition_utils-mesh_utils.md
- T5x Partitioning: generated-partition_utils-t5x_partitioning.md
- Utils: generated-utils.md
- Xrapture:
- Implicit Array: generated-xrapture-implicit_array.md
- Tracer: generated-xrapture-tracer.md
- Xrapture: generated-xrapture-xrapture.md
plugins:
- search
- mkdocstrings:
handlers:
python:
options:
docstring_style: sphinx
repo_url: https://github.com/erfanzar/FJFormer
site_author: Erfan Zare Chavoshi
site_name: FJFormer
copyright: Erfan Zare Chavoshi-FJFormer
theme:
highlightjs: true
hljs_languages:
- yaml
- python
name: material