Skip to content

Commit c268eac

Browse files
Introduce T5Gemma model presets in KerasHub (#2373)
* init: Add model presets * feat: Update T5Gemma model parameters and Kaggle handles in presets
1 parent 914e730 commit c268eac

File tree

1 file changed

+360
-1
lines changed

1 file changed

+360
-1
lines changed

keras_hub/src/models/t5gemma/t5gemma_presets.py

Lines changed: 360 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,85 @@
11
# Metadata for loading pretrained model weights.
22
backbone_presets = {
3+
"t5gemma_s_s_ul2": {
4+
"metadata": {
5+
"description": (
6+
"T5Gemma S/S model with a small encoder and small decoder, "
7+
"adapted as a UL2 model."
8+
),
9+
"params": 312517632,
10+
"path": "t5gemma",
11+
},
12+
"kaggle_handle": "kaggle://keras/t5-gemma/keras/t5gemma_s_s_ul2/1",
13+
},
14+
"t5gemma_s_s_prefixlm": {
15+
"metadata": {
16+
"description": (
17+
"T5Gemma S/S model with a small encoder and small decoder, "
18+
"adapted as a prefix language model."
19+
),
20+
"params": 312517632,
21+
"path": "t5gemma",
22+
},
23+
"kaggle_handle": "kaggle://keras/t5-gemma/keras/t5gemma_s_s_prefixlm/1",
24+
},
25+
"t5gemma_s_s_ul2_it": {
26+
"metadata": {
27+
"description": (
28+
"T5Gemma S/S model with a small encoder and small decoder, "
29+
"adapted as a UL2 model and fine-tuned for instruction "
30+
"following."
31+
),
32+
"params": 312517632,
33+
"path": "t5gemma",
34+
},
35+
"kaggle_handle": "kaggle://keras/t5-gemma/keras/t5gemma_s_s_ul2_it/1",
36+
},
37+
"t5gemma_s_s_prefixlm_it": {
38+
"metadata": {
39+
"description": (
40+
"T5Gemma S/S model with a small encoder and small decoder, "
41+
"adapted as a prefix language model and fine-tuned for "
42+
"instruction following."
43+
),
44+
"params": 312517632,
45+
"path": "t5gemma",
46+
},
47+
"kaggle_handle": "kaggle://keras/t5-gemma/keras/t5gemma_s_s_prefixlm_it/1",
48+
},
49+
"t5gemma_b_b_ul2": {
50+
"metadata": {
51+
"description": (
52+
"T5Gemma B/B model with a base encoder and base decoder, "
53+
"adapted as a UL2 model."
54+
),
55+
"params": 591490560,
56+
"path": "t5gemma",
57+
},
58+
"kaggle_handle": "kaggle://keras/t5-gemma/keras/t5gemma_b_b_ul2/1",
59+
},
60+
"t5gemma_b_b_prefixlm": {
61+
"metadata": {
62+
"description": (
63+
"T5Gemma B/B model with a base encoder and base decoder, "
64+
"adapted as a prefix language model."
65+
),
66+
"params": 591490560,
67+
"path": "t5gemma",
68+
},
69+
"kaggle_handle": "kaggle://keras/t5-gemma/keras/t5gemma_b_b_prefixlm/1",
70+
},
71+
"t5gemma_b_b_ul2_it": {
72+
"metadata": {
73+
"description": (
74+
"T5Gemma B/B model with a base encoder and base decoder, "
75+
"adapted as a UL2 model and fine-tuned for instruction "
76+
"following."
77+
),
78+
"params": 591490560,
79+
"path": "t5gemma",
80+
},
81+
"kaggle_handle": "kaggle://keras/t5-gemma/keras/t5gemma_b_b_ul2_it/1",
82+
},
383
"t5gemma_b_b_prefixlm_it": {
484
"metadata": {
585
"description": (
@@ -10,6 +90,285 @@
1090
"params": 591490560,
1191
"path": "t5gemma",
1292
},
13-
"kaggle_handle": "kaggle://harshaljanjani/t5gemma/keras/t5gemma_b_b_prefixlm_it",
93+
"kaggle_handle": "kaggle://keras/t5-gemma/keras/t5gemma_b_b_prefixlm_it/1",
94+
},
95+
"t5gemma_l_l_ul2": {
96+
"metadata": {
97+
"description": (
98+
"T5Gemma L/L model with a large encoder and large decoder, "
99+
"adapted as a UL2 model."
100+
),
101+
"params": 1241761792,
102+
"path": "t5gemma",
103+
},
104+
"kaggle_handle": "kaggle://keras/t5-gemma/keras/t5gemma_l_l_ul2/1",
105+
},
106+
"t5gemma_l_l_prefixlm": {
107+
"metadata": {
108+
"description": (
109+
"T5Gemma L/L model with a large encoder and large decoder, "
110+
"adapted as a prefix language model."
111+
),
112+
"params": 1241761792,
113+
"path": "t5gemma",
114+
},
115+
"kaggle_handle": "kaggle://keras/t5-gemma/keras/t5gemma_l_l_prefixlm/1",
116+
},
117+
"t5gemma_l_l_ul2_it": {
118+
"metadata": {
119+
"description": (
120+
"T5Gemma L/L model with a large encoder and large decoder, "
121+
"adapted as a UL2 model and fine-tuned for instruction "
122+
"following."
123+
),
124+
"params": 1241761792,
125+
"path": "t5gemma",
126+
},
127+
"kaggle_handle": "kaggle://keras/t5-gemma/keras/t5gemma_l_l_ul2_it/1",
128+
},
129+
"t5gemma_l_l_prefixlm_it": {
130+
"metadata": {
131+
"description": (
132+
"T5Gemma L/L model with a large encoder and large decoder, "
133+
"adapted as a prefix language model and fine-tuned for "
134+
"instruction following."
135+
),
136+
"params": 1241761792,
137+
"path": "t5gemma",
138+
},
139+
"kaggle_handle": "kaggle://keras/t5-gemma/keras/t5gemma_l_l_prefixlm_it/1",
140+
},
141+
"t5gemma_ml_ml_ul2": {
142+
"metadata": {
143+
"description": (
144+
"T5Gemma ML/ML model with a medium-large encoder and "
145+
"medium-large decoder, adapted as a UL2 model."
146+
),
147+
"params": 2200345344,
148+
"path": "t5gemma",
149+
},
150+
"kaggle_handle": "kaggle://keras/t5-gemma/keras/t5gemma_ml_ml_ul2/1",
151+
},
152+
"t5gemma_ml_ml_prefixlm": {
153+
"metadata": {
154+
"description": (
155+
"T5Gemma ML/ML model with a medium-large encoder and "
156+
"medium-large decoder, adapted as a prefix language model."
157+
),
158+
"params": 2200345344,
159+
"path": "t5gemma",
160+
},
161+
"kaggle_handle": "kaggle://keras/t5-gemma/keras/t5gemma_ml_ml_prefixlm/1",
162+
},
163+
"t5gemma_ml_ml_ul2_it": {
164+
"metadata": {
165+
"description": (
166+
"T5Gemma ML/ML model with a medium-large encoder and "
167+
"medium-large decoder, adapted as a UL2 model and fine-tuned "
168+
"for instruction following."
169+
),
170+
"params": 2200345344,
171+
"path": "t5gemma",
172+
},
173+
"kaggle_handle": "kaggle://keras/t5-gemma/keras/t5gemma_ml_ml_ul2_it/1",
174+
},
175+
"t5gemma_ml_ml_prefixlm_it": {
176+
"metadata": {
177+
"description": (
178+
"T5Gemma ML/ML model with a medium-large encoder and "
179+
"medium-large decoder, adapted as a prefix language model and "
180+
"fine-tuned for instruction following."
181+
),
182+
"params": 2200345344,
183+
"path": "t5gemma",
184+
},
185+
"kaggle_handle": "kaggle://keras/t5-gemma/keras/t5gemma_ml_ml_prefixlm_it/1",
186+
},
187+
"t5gemma_xl_xl_ul2": {
188+
"metadata": {
189+
"description": (
190+
"T5Gemma XL/XL model with an extra-large encoder and "
191+
"extra-large decoder, adapted as a UL2 model."
192+
),
193+
"params": 3766980608,
194+
"path": "t5gemma",
195+
},
196+
"kaggle_handle": "kaggle://keras/t5-gemma/keras/t5gemma_xl_xl_ul2/1",
197+
},
198+
"t5gemma_xl_xl_prefixlm": {
199+
"metadata": {
200+
"description": (
201+
"T5Gemma XL/XL model with an extra-large encoder and "
202+
"extra-large decoder, adapted as a prefix language model."
203+
),
204+
"params": 3766980608,
205+
"path": "t5gemma",
206+
},
207+
"kaggle_handle": "kaggle://keras/t5-gemma/keras/t5gemma_xl_xl_prefixlm/1",
208+
},
209+
"t5gemma_xl_xl_ul2_it": {
210+
"metadata": {
211+
"description": (
212+
"T5Gemma XL/XL model with an extra-large encoder and "
213+
"extra-large decoder, adapted as a UL2 model and fine-tuned "
214+
"for instruction following."
215+
),
216+
"params": 3766980608,
217+
"path": "t5gemma",
218+
},
219+
"kaggle_handle": "kaggle://keras/t5-gemma/keras/t5gemma_xl_xl_ul2_it/1",
220+
},
221+
"t5gemma_xl_xl_prefixlm_it": {
222+
"metadata": {
223+
"description": (
224+
"T5Gemma XL/XL model with an extra-large encoder and "
225+
"extra-large decoder, adapted as a prefix language model and "
226+
"fine-tuned for instruction following."
227+
),
228+
"params": 3766980608,
229+
"path": "t5gemma",
230+
},
231+
"kaggle_handle": "kaggle://keras/t5-gemma/keras/t5gemma_xl_xl_prefixlm_it/1",
232+
},
233+
"t5gemma_2b_2b_ul2": {
234+
"metadata": {
235+
"description": (
236+
"T5Gemma 2B/2B model with a 2-billion-parameter encoder and "
237+
"2-billion-parameter decoder, adapted as a UL2 model."
238+
),
239+
"params": 5596853760,
240+
"path": "t5gemma",
241+
},
242+
"kaggle_handle": "kaggle://keras/t5-gemma/keras/t5gemma_2b_2b_ul2/1",
243+
},
244+
"t5gemma_2b_2b_prefixlm": {
245+
"metadata": {
246+
"description": (
247+
"T5Gemma 2B/2B model with a 2-billion-parameter encoder and "
248+
"2-billion-parameter decoder, adapted as a prefix language "
249+
"model."
250+
),
251+
"params": 5596853760,
252+
"path": "t5gemma",
253+
},
254+
"kaggle_handle": "kaggle://keras/t5-gemma/keras/t5gemma_2b_2b_prefixlm/1",
255+
},
256+
"t5gemma_2b_2b_ul2_it": {
257+
"metadata": {
258+
"description": (
259+
"T5Gemma 2B/2B model with a 2-billion-parameter encoder and "
260+
"2-billion-parameter decoder, adapted as a UL2 model and "
261+
"fine-tuned for instruction following."
262+
),
263+
"params": 5596853760,
264+
"path": "t5gemma",
265+
},
266+
"kaggle_handle": "kaggle://keras/t5-gemma/keras/t5gemma_2b_2b_ul2_it/1",
267+
},
268+
"t5gemma_2b_2b_prefixlm_it": {
269+
"metadata": {
270+
"description": (
271+
"T5Gemma 2B/2B model with a 2-billion-parameter encoder and "
272+
"2-billion-parameter decoder, adapted as a prefix language "
273+
"model and fine-tuned for instruction following."
274+
),
275+
"params": 5596853760,
276+
"path": "t5gemma",
277+
},
278+
"kaggle_handle": "kaggle://keras/t5-gemma/keras/t5gemma_2b_2b_prefixlm_it/1",
279+
},
280+
"t5gemma_9b_2b_ul2": {
281+
"metadata": {
282+
"description": (
283+
"T5Gemma 9B/2B model with a 9-billion-parameter encoder and "
284+
"2-billion-parameter decoder, adapted as a UL2 model."
285+
),
286+
"params": 12292375296,
287+
"path": "t5gemma",
288+
},
289+
"kaggle_handle": "kaggle://keras/t5-gemma/keras/t5gemma_9b_2b_ul2/1",
290+
},
291+
"t5gemma_9b_2b_prefixlm": {
292+
"metadata": {
293+
"description": (
294+
"T5Gemma 9B/2B model with a 9-billion-parameter encoder and "
295+
"2-billion-parameter decoder, adapted as a prefix language "
296+
"model."
297+
),
298+
"params": 12292375296,
299+
"path": "t5gemma",
300+
},
301+
"kaggle_handle": "kaggle://keras/t5-gemma/keras/t5gemma_9b_2b_prefixlm/1",
302+
},
303+
"t5gemma_9b_2b_ul2_it": {
304+
"metadata": {
305+
"description": (
306+
"T5Gemma 9B/2B model with a 9-billion-parameter encoder and "
307+
"2-billion-parameter decoder, adapted as a UL2 model and "
308+
"fine-tuned for instruction following."
309+
),
310+
"params": 12292375296,
311+
"path": "t5gemma",
312+
},
313+
"kaggle_handle": "kaggle://keras/t5-gemma/keras/t5gemma_9b_2b_ul2_it/1",
314+
},
315+
"t5gemma_9b_2b_prefixlm_it": {
316+
"metadata": {
317+
"description": (
318+
"T5Gemma 9B/2B model with a 9-billion-parameter encoder and "
319+
"2-billion-parameter decoder, adapted as a prefix language "
320+
"model and fine-tuned for instruction following."
321+
),
322+
"params": 12292375296,
323+
"path": "t5gemma",
324+
},
325+
"kaggle_handle": "kaggle://keras/t5-gemma/keras/t5gemma_9b_2b_prefixlm_it/1",
326+
},
327+
"t5gemma_9b_9b_ul2": {
328+
"metadata": {
329+
"description": (
330+
"T5Gemma 9B/9B model with a 9-billion-parameter encoder and "
331+
"9-billion-parameter decoder, adapted as a UL2 model."
332+
),
333+
"params": 20333401088,
334+
"path": "t5gemma",
335+
},
336+
"kaggle_handle": "kaggle://keras/t5-gemma/keras/t5gemma_9b_9b_ul2/1",
337+
},
338+
"t5gemma_9b_9b_prefixlm": {
339+
"metadata": {
340+
"description": (
341+
"T5Gemma 9B/9B model with a 9-billion-parameter encoder and "
342+
"9-billion-parameter decoder, adapted as a prefix language "
343+
"model."
344+
),
345+
"params": 20333401088,
346+
"path": "t5gemma",
347+
},
348+
"kaggle_handle": "kaggle://keras/t5-gemma/keras/t5gemma_9b_9b_prefixlm/1",
349+
},
350+
"t5gemma_9b_9b_ul2_it": {
351+
"metadata": {
352+
"description": (
353+
"T5Gemma 9B/9B model with a 9-billion-parameter encoder and "
354+
"9-billion-parameter decoder, adapted as a UL2 model and "
355+
"fine-tuned for instruction following."
356+
),
357+
"params": 20333401088,
358+
"path": "t5gemma",
359+
},
360+
"kaggle_handle": "kaggle://keras/t5-gemma/keras/t5gemma_9b_9b_ul2_it/1",
361+
},
362+
"t5gemma_9b_9b_prefixlm_it": {
363+
"metadata": {
364+
"description": (
365+
"T5Gemma 9B/9B model with a 9-billion-parameter encoder and "
366+
"9-billion-parameter decoder, adapted as a prefix language "
367+
"model and fine-tuned for instruction following."
368+
),
369+
"params": 20333401088,
370+
"path": "t5gemma",
371+
},
372+
"kaggle_handle": "kaggle://keras/t5-gemma/keras/t5gemma_9b_9b_prefixlm_it/1",
14373
},
15374
}

0 commit comments

Comments
 (0)