Commit History

another test
d09d63c

mehdi999 commited on

Changed requirements
78e8994

mehdi999 commited on

Corrected pytorch version
39d8b01

mehdi999 commited on

fix(glA): init default conv_state at first prefill to avoid NoneType unpack
72dc299

mehdi999 commited on

fix: remove leftover cache=cache in text_to_speech call
677e61f

mehdi999 commited on

fix: let decoder manage its own cache (avoid None conv_state in FLA)
750aa27

mehdi999 commited on

back to basics
fd1f480

mehdi999 commited on

runtime: set ZeroGPU duration to 600s
6d19e74

mehdi999 commited on

runtime: set ZeroGPU duration to 600s
158b7a6

mehdi999 commited on

Space: preload CPU thread + cache + logs
6d29905

mehdi999 commited on

added few things
92ec5fe

mehdi999 commited on

added few things
58c000c

mehdi999 commited on

added few things
313c415

mehdi999 commited on

added few things
5edbebd

mehdi999 commited on

runtime: set ZeroGPU duration to 600s
018aa79

mehdi999 commited on

added few things
fc1f9f2

mehdi999 commited on

added few things
9f2e2fc

mehdi999 commited on

added few things
2dc4aff

mehdi999 commited on

added few things
5997b2e

mehdi999 commited on

Fix: safe FLA backend + cache guards
cf9d6a9

mehdi999 commited on

added few things
831395c

mehdi999 commited on

added few things
55fecbf

mehdi999 commited on

added few things
a7c605a

mehdi999 commited on

added few things
f6f4ba0

mehdi999 commited on

added watch
6b8706f

mehdi999 commited on

added watch
3d734f0

mehdi999 commited on

runtime: set ZeroGPU duration to 600s
7208504

mehdi999 commited on

chore: retrigger build
dd090c8

mehdi999 commited on

runtime: extend ZeroGPU duration to 600s
dbbf6f4

mehdi999 commited on

fix: pass num_heads to CrossAttention ctor
46e132f

mehdi999 commited on

deps: drop causal-conv1d (needs torch at build-time; not required for demo)
5fd5d30

mehdi999 commited on

fix: SwiGLU ctor signature (drop out_features); add causal-conv1d
7314545

mehdi999 commited on

fix: expose SimpleGLADecoder.config for registry instantiate_from_config
ca4ede6

mehdi999 commited on

Force FLA mode=chunk to avoid Triton fused kernels on ZeroGPU
b90b280

mehdi999 commited on

chore: retrigger build
50be982

mehdi999 commited on

Force FLA mode=chunk to avoid Triton fused kernels on ZeroGPU
a175cfa

mehdi999 commited on

fix(GLA): provide safe past_key_values with default conv_state=(None,None,None)
bb32e4f

mehdi999 commited on

fix(GLA): provide safe past_key_values with default conv_state=(None,None,None)
80ec7ec

mehdi999 commited on

fix(GLA): provide safe past_key_values with default conv_state=(None,None,None)
a69b69a

mehdi999 commited on

fix(GLA): provide safe past_key_values with default conv_state=(None,None,None)
159df2e

mehdi999 commited on

fix: remove leftover cache=cache from text_to_speech call
8caf57f

mehdi999 commited on

inference: remove empty cache (GLA expects prefetched conv_state)
d7cfd9b

mehdi999 commited on

chore: retrigger build
ad26fe4

mehdi999 commited on

fix: handle VelocityHeadSamplingParams signature; better error logging
0a0019f

mehdi999 commited on

fix: align torch/torchaudio to cu124 (CUDA 12.4)
1075fc0

mehdi999 commited on

deps: add flash-linear-attention; bump torch/torchaudio to 2.5.1 and transformers to 4.53.0
c99b6f0

mehdi999 commited on

fix: replace zcodec import with relative (TransformerBlock)
b8aa630

mehdi999 commited on

inference-only: drop zcodec deps (relative imports) and avoid training/tokenizer imports at package init
b650fff

mehdi999 commited on

deps: drop flash-linear-attention to resolve Torch/Transformers conflicts
0d64ef9

mehdi999 commited on

deps: align with pyproject; torch/torchaudio pinned to cu121-compatible
6a0a374

mehdi999 commited on