model_prices_and_context_window.json 25 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717718719720721722723724725726727728729730731732733734735736737738739740741742743744745746747748749750751752753754755756757758759760761762763764765766767768769770771772773774775776777778779780781782783784785786787788789790791792793794795796797798799800801802803
  1. {
  2. "openai/gpt-4": {
  3. "max_tokens": 4096,
  4. "max_input_tokens": 8192,
  5. "max_output_tokens": 4096,
  6. "input_cost_per_token": 0.00003,
  7. "output_cost_per_token": 0.00006
  8. },
  9. "openai/gpt-4o": {
  10. "max_tokens": 4096,
  11. "max_input_tokens": 128000,
  12. "max_output_tokens": 4096,
  13. "input_cost_per_token": 0.000005,
  14. "output_cost_per_token": 0.000015
  15. },
  16. "openai/gpt-4o-2024-05-13": {
  17. "max_tokens": 4096,
  18. "max_input_tokens": 128000,
  19. "max_output_tokens": 4096,
  20. "input_cost_per_token": 0.000005,
  21. "output_cost_per_token": 0.000015
  22. },
  23. "openai/gpt-4-turbo-preview": {
  24. "max_tokens": 4096,
  25. "max_input_tokens": 128000,
  26. "max_output_tokens": 4096,
  27. "input_cost_per_token": 0.00001,
  28. "output_cost_per_token": 0.00003
  29. },
  30. "openai/gpt-4-0314": {
  31. "max_tokens": 4096,
  32. "max_input_tokens": 8192,
  33. "max_output_tokens": 4096,
  34. "input_cost_per_token": 0.00003,
  35. "output_cost_per_token": 0.00006
  36. },
  37. "openai/gpt-4-0613": {
  38. "max_tokens": 4096,
  39. "max_input_tokens": 8192,
  40. "max_output_tokens": 4096,
  41. "input_cost_per_token": 0.00003,
  42. "output_cost_per_token": 0.00006
  43. },
  44. "openai/gpt-4-32k": {
  45. "max_tokens": 4096,
  46. "max_input_tokens": 32768,
  47. "max_output_tokens": 4096,
  48. "input_cost_per_token": 0.00006,
  49. "output_cost_per_token": 0.00012
  50. },
  51. "openai/gpt-4-32k-0314": {
  52. "max_tokens": 4096,
  53. "max_input_tokens": 32768,
  54. "max_output_tokens": 4096,
  55. "input_cost_per_token": 0.00006,
  56. "output_cost_per_token": 0.00012
  57. },
  58. "openai/gpt-4-32k-0613": {
  59. "max_tokens": 4096,
  60. "max_input_tokens": 32768,
  61. "max_output_tokens": 4096,
  62. "input_cost_per_token": 0.00006,
  63. "output_cost_per_token": 0.00012
  64. },
  65. "openai/gpt-4-turbo": {
  66. "max_tokens": 4096,
  67. "max_input_tokens": 128000,
  68. "max_output_tokens": 4096,
  69. "input_cost_per_token": 0.00001,
  70. "output_cost_per_token": 0.00003
  71. },
  72. "openai/gpt-4-turbo-2024-04-09": {
  73. "max_tokens": 4096,
  74. "max_input_tokens": 128000,
  75. "max_output_tokens": 4096,
  76. "input_cost_per_token": 0.00001,
  77. "output_cost_per_token": 0.00003
  78. },
  79. "openai/gpt-4-1106-preview": {
  80. "max_tokens": 4096,
  81. "max_input_tokens": 128000,
  82. "max_output_tokens": 4096,
  83. "input_cost_per_token": 0.00001,
  84. "output_cost_per_token": 0.00003
  85. },
  86. "openai/gpt-4-0125-preview": {
  87. "max_tokens": 4096,
  88. "max_input_tokens": 128000,
  89. "max_output_tokens": 4096,
  90. "input_cost_per_token": 0.00001,
  91. "output_cost_per_token": 0.00003
  92. },
  93. "openai/gpt-3.5-turbo": {
  94. "max_tokens": 4097,
  95. "max_input_tokens": 16385,
  96. "max_output_tokens": 4096,
  97. "input_cost_per_token": 0.0000015,
  98. "output_cost_per_token": 0.000002
  99. },
  100. "openai/gpt-3.5-turbo-0301": {
  101. "max_tokens": 4097,
  102. "max_input_tokens": 4097,
  103. "max_output_tokens": 4096,
  104. "input_cost_per_token": 0.0000015,
  105. "output_cost_per_token": 0.000002
  106. },
  107. "openai/gpt-3.5-turbo-0613": {
  108. "input_cost_per_token": 0.0000015,
  109. "output_cost_per_token": 0.000002
  110. },
  111. "openai/gpt-3.5-turbo-1106": {
  112. "max_tokens": 16385,
  113. "max_input_tokens": 16385,
  114. "max_output_tokens": 4096,
  115. "input_cost_per_token": 0.0000010,
  116. "output_cost_per_token": 0.0000020
  117. },
  118. "openai/gpt-3.5-turbo-0125": {
  119. "max_tokens": 16385,
  120. "max_input_tokens": 16385,
  121. "max_output_tokens": 4096,
  122. "input_cost_per_token": 0.0000005,
  123. "output_cost_per_token": 0.0000015
  124. },
  125. "openai/gpt-3.5-turbo-16k": {
  126. "max_tokens": 16385,
  127. "max_input_tokens": 16385,
  128. "max_output_tokens": 4096,
  129. "input_cost_per_token": 0.000003,
  130. "output_cost_per_token": 0.000004
  131. },
  132. "openai/gpt-3.5-turbo-16k-0613": {
  133. "max_tokens": 16385,
  134. "max_input_tokens": 16385,
  135. "max_output_tokens": 4096,
  136. "input_cost_per_token": 0.000003,
  137. "output_cost_per_token": 0.000004
  138. },
  139. "openai/text-embedding-3-large": {
  140. "max_tokens": 8191,
  141. "max_input_tokens": 8191,
  142. "output_vector_size": 3072,
  143. "input_cost_per_token": 0.00000013,
  144. "output_cost_per_token": 0.000000
  145. },
  146. "openai/text-embedding-3-small": {
  147. "max_tokens": 8191,
  148. "max_input_tokens": 8191,
  149. "output_vector_size": 1536,
  150. "input_cost_per_token": 0.00000002,
  151. "output_cost_per_token": 0.000000
  152. },
  153. "openai/text-embedding-ada-002": {
  154. "max_tokens": 8191,
  155. "max_input_tokens": 8191,
  156. "output_vector_size": 1536,
  157. "input_cost_per_token": 0.0000001,
  158. "output_cost_per_token": 0.000000
  159. },
  160. "openai/text-embedding-ada-002-v2": {
  161. "max_tokens": 8191,
  162. "max_input_tokens": 8191,
  163. "input_cost_per_token": 0.0000001,
  164. "output_cost_per_token": 0.000000
  165. },
  166. "openai/babbage-002": {
  167. "max_tokens": 16384,
  168. "max_input_tokens": 16384,
  169. "max_output_tokens": 4096,
  170. "input_cost_per_token": 0.0000004,
  171. "output_cost_per_token": 0.0000004
  172. },
  173. "openai/davinci-002": {
  174. "max_tokens": 16384,
  175. "max_input_tokens": 16384,
  176. "max_output_tokens": 4096,
  177. "input_cost_per_token": 0.000002,
  178. "output_cost_per_token": 0.000002
  179. },
  180. "openai/gpt-3.5-turbo-instruct": {
  181. "max_tokens": 4096,
  182. "max_input_tokens": 8192,
  183. "max_output_tokens": 4096,
  184. "input_cost_per_token": 0.0000015,
  185. "output_cost_per_token": 0.000002
  186. },
  187. "openai/gpt-3.5-turbo-instruct-0914": {
  188. "max_tokens": 4097,
  189. "max_input_tokens": 8192,
  190. "max_output_tokens": 4097,
  191. "input_cost_per_token": 0.0000015,
  192. "output_cost_per_token": 0.000002
  193. },
  194. "azure/gpt-4o": {
  195. "max_tokens": 4096,
  196. "max_input_tokens": 128000,
  197. "max_output_tokens": 4096,
  198. "input_cost_per_token": 0.000005,
  199. "output_cost_per_token": 0.000015
  200. },
  201. "azure/gpt-4-turbo-2024-04-09": {
  202. "max_tokens": 4096,
  203. "max_input_tokens": 128000,
  204. "max_output_tokens": 4096,
  205. "input_cost_per_token": 0.00001,
  206. "output_cost_per_token": 0.00003
  207. },
  208. "azure/gpt-4-0125-preview": {
  209. "max_tokens": 4096,
  210. "max_input_tokens": 128000,
  211. "max_output_tokens": 4096,
  212. "input_cost_per_token": 0.00001,
  213. "output_cost_per_token": 0.00003
  214. },
  215. "azure/gpt-4-1106-preview": {
  216. "max_tokens": 4096,
  217. "max_input_tokens": 128000,
  218. "max_output_tokens": 4096,
  219. "input_cost_per_token": 0.00001,
  220. "output_cost_per_token": 0.00003
  221. },
  222. "azure/gpt-4-0613": {
  223. "max_tokens": 4096,
  224. "max_input_tokens": 8192,
  225. "max_output_tokens": 4096,
  226. "input_cost_per_token": 0.00003,
  227. "output_cost_per_token": 0.00006
  228. },
  229. "azure/gpt-4-32k-0613": {
  230. "max_tokens": 4096,
  231. "max_input_tokens": 32768,
  232. "max_output_tokens": 4096,
  233. "input_cost_per_token": 0.00006,
  234. "output_cost_per_token": 0.00012
  235. },
  236. "azure/gpt-4-32k": {
  237. "max_tokens": 4096,
  238. "max_input_tokens": 32768,
  239. "max_output_tokens": 4096,
  240. "input_cost_per_token": 0.00006,
  241. "output_cost_per_token": 0.00012
  242. },
  243. "azure/gpt-4": {
  244. "max_tokens": 4096,
  245. "max_input_tokens": 8192,
  246. "max_output_tokens": 4096,
  247. "input_cost_per_token": 0.00003,
  248. "output_cost_per_token": 0.00006
  249. },
  250. "azure/gpt-4-turbo": {
  251. "max_tokens": 4096,
  252. "max_input_tokens": 128000,
  253. "max_output_tokens": 4096,
  254. "input_cost_per_token": 0.00001,
  255. "output_cost_per_token": 0.00003
  256. },
  257. "azure/gpt-4-turbo-vision-preview": {
  258. "max_tokens": 4096,
  259. "max_input_tokens": 128000,
  260. "max_output_tokens": 4096,
  261. "input_cost_per_token": 0.00001,
  262. "output_cost_per_token": 0.00003
  263. },
  264. "azure/gpt-3.5-turbo-16k-0613": {
  265. "max_tokens": 4096,
  266. "max_input_tokens": 16385,
  267. "max_output_tokens": 4096,
  268. "input_cost_per_token": 0.000003,
  269. "output_cost_per_token": 0.000004
  270. },
  271. "azure/gpt-3.5-turbo-1106": {
  272. "max_tokens": 4096,
  273. "max_input_tokens": 16384,
  274. "max_output_tokens": 4096,
  275. "input_cost_per_token": 0.0000015,
  276. "output_cost_per_token": 0.000002
  277. },
  278. "azure/gpt-3.5-turbo-0125": {
  279. "max_tokens": 4096,
  280. "max_input_tokens": 16384,
  281. "max_output_tokens": 4096,
  282. "input_cost_per_token": 0.0000005,
  283. "output_cost_per_token": 0.0000015
  284. },
  285. "azure/gpt-3.5-turbo-16k": {
  286. "max_tokens": 4096,
  287. "max_input_tokens": 16385,
  288. "max_output_tokens": 4096,
  289. "input_cost_per_token": 0.000003,
  290. "output_cost_per_token": 0.000004
  291. },
  292. "azure/gpt-3.5-turbo": {
  293. "max_tokens": 4096,
  294. "max_input_tokens": 4097,
  295. "max_output_tokens": 4096,
  296. "input_cost_per_token": 0.0000005,
  297. "output_cost_per_token": 0.0000015
  298. },
  299. "azure/gpt-3.5-turbo-instruct-0914": {
  300. "max_tokens": 4097,
  301. "max_input_tokens": 4097,
  302. "input_cost_per_token": 0.0000015,
  303. "output_cost_per_token": 0.000002
  304. },
  305. "azure/gpt-3.5-turbo-instruct": {
  306. "max_tokens": 4097,
  307. "max_input_tokens": 4097,
  308. "input_cost_per_token": 0.0000015,
  309. "output_cost_per_token": 0.000002
  310. },
  311. "azure/text-embedding-ada-002": {
  312. "max_tokens": 8191,
  313. "max_input_tokens": 8191,
  314. "input_cost_per_token": 0.0000001,
  315. "output_cost_per_token": 0.000000
  316. },
  317. "azure/text-embedding-3-large": {
  318. "max_tokens": 8191,
  319. "max_input_tokens": 8191,
  320. "input_cost_per_token": 0.00000013,
  321. "output_cost_per_token": 0.000000
  322. },
  323. "azure/text-embedding-3-small": {
  324. "max_tokens": 8191,
  325. "max_input_tokens": 8191,
  326. "input_cost_per_token": 0.00000002,
  327. "output_cost_per_token": 0.000000
  328. },
  329. "mistralai/mistral-tiny": {
  330. "max_tokens": 8191,
  331. "max_input_tokens": 32000,
  332. "max_output_tokens": 8191,
  333. "input_cost_per_token": 0.00000025,
  334. "output_cost_per_token": 0.00000025
  335. },
  336. "mistralai/mistral-small": {
  337. "max_tokens": 8191,
  338. "max_input_tokens": 32000,
  339. "max_output_tokens": 8191,
  340. "input_cost_per_token": 0.000001,
  341. "output_cost_per_token": 0.000003
  342. },
  343. "mistralai/mistral-small-latest": {
  344. "max_tokens": 8191,
  345. "max_input_tokens": 32000,
  346. "max_output_tokens": 8191,
  347. "input_cost_per_token": 0.000001,
  348. "output_cost_per_token": 0.000003
  349. },
  350. "mistralai/mistral-medium": {
  351. "max_tokens": 8191,
  352. "max_input_tokens": 32000,
  353. "max_output_tokens": 8191,
  354. "input_cost_per_token": 0.0000027,
  355. "output_cost_per_token": 0.0000081
  356. },
  357. "mistralai/mistral-medium-latest": {
  358. "max_tokens": 8191,
  359. "max_input_tokens": 32000,
  360. "max_output_tokens": 8191,
  361. "input_cost_per_token": 0.0000027,
  362. "output_cost_per_token": 0.0000081
  363. },
  364. "mistralai/mistral-medium-2312": {
  365. "max_tokens": 8191,
  366. "max_input_tokens": 32000,
  367. "max_output_tokens": 8191,
  368. "input_cost_per_token": 0.0000027,
  369. "output_cost_per_token": 0.0000081
  370. },
  371. "mistralai/mistral-large-latest": {
  372. "max_tokens": 8191,
  373. "max_input_tokens": 32000,
  374. "max_output_tokens": 8191,
  375. "input_cost_per_token": 0.000004,
  376. "output_cost_per_token": 0.000012
  377. },
  378. "mistralai/mistral-large-2402": {
  379. "max_tokens": 8191,
  380. "max_input_tokens": 32000,
  381. "max_output_tokens": 8191,
  382. "input_cost_per_token": 0.000004,
  383. "output_cost_per_token": 0.000012
  384. },
  385. "mistralai/open-mistral-7b": {
  386. "max_tokens": 8191,
  387. "max_input_tokens": 32000,
  388. "max_output_tokens": 8191,
  389. "input_cost_per_token": 0.00000025,
  390. "output_cost_per_token": 0.00000025
  391. },
  392. "mistralai/open-mixtral-8x7b": {
  393. "max_tokens": 8191,
  394. "max_input_tokens": 32000,
  395. "max_output_tokens": 8191,
  396. "input_cost_per_token": 0.0000007,
  397. "output_cost_per_token": 0.0000007
  398. },
  399. "mistralai/open-mixtral-8x22b": {
  400. "max_tokens": 8191,
  401. "max_input_tokens": 64000,
  402. "max_output_tokens": 8191,
  403. "input_cost_per_token": 0.000002,
  404. "output_cost_per_token": 0.000006
  405. },
  406. "mistralai/codestral-latest": {
  407. "max_tokens": 8191,
  408. "max_input_tokens": 32000,
  409. "max_output_tokens": 8191,
  410. "input_cost_per_token": 0.000001,
  411. "output_cost_per_token": 0.000003
  412. },
  413. "mistralai/codestral-2405": {
  414. "max_tokens": 8191,
  415. "max_input_tokens": 32000,
  416. "max_output_tokens": 8191,
  417. "input_cost_per_token": 0.000001,
  418. "output_cost_per_token": 0.000003
  419. },
  420. "mistralai/mistral-embed": {
  421. "max_tokens": 8192,
  422. "max_input_tokens": 8192,
  423. "input_cost_per_token": 0.0000001,
  424. "output_cost_per_token": 0.0
  425. },
  426. "groq/llama2-70b-4096": {
  427. "max_tokens": 4096,
  428. "max_input_tokens": 4096,
  429. "max_output_tokens": 4096,
  430. "input_cost_per_token": 0.00000070,
  431. "output_cost_per_token": 0.00000080
  432. },
  433. "groq/llama3-8b-8192": {
  434. "max_tokens": 8192,
  435. "max_input_tokens": 8192,
  436. "max_output_tokens": 8192,
  437. "input_cost_per_token": 0.00000010,
  438. "output_cost_per_token": 0.00000010
  439. },
  440. "groq/llama3-70b-8192": {
  441. "max_tokens": 8192,
  442. "max_input_tokens": 8192,
  443. "max_output_tokens": 8192,
  444. "input_cost_per_token": 0.00000064,
  445. "output_cost_per_token": 0.00000080
  446. },
  447. "groq/mixtral-8x7b-32768": {
  448. "max_tokens": 32768,
  449. "max_input_tokens": 32768,
  450. "max_output_tokens": 32768,
  451. "input_cost_per_token": 0.00000027,
  452. "output_cost_per_token": 0.00000027
  453. },
  454. "groq/gemma-7b-it": {
  455. "max_tokens": 8192,
  456. "max_input_tokens": 8192,
  457. "max_output_tokens": 8192,
  458. "input_cost_per_token": 0.00000010,
  459. "output_cost_per_token": 0.00000010
  460. },
  461. "anthropic/claude-instant-1": {
  462. "max_tokens": 8191,
  463. "max_input_tokens": 100000,
  464. "max_output_tokens": 8191,
  465. "input_cost_per_token": 0.00000163,
  466. "output_cost_per_token": 0.00000551
  467. },
  468. "anthropic/claude-instant-1.2": {
  469. "max_tokens": 8191,
  470. "max_input_tokens": 100000,
  471. "max_output_tokens": 8191,
  472. "input_cost_per_token": 0.000000163,
  473. "output_cost_per_token": 0.000000551
  474. },
  475. "anthropic/claude-2": {
  476. "max_tokens": 8191,
  477. "max_input_tokens": 100000,
  478. "max_output_tokens": 8191,
  479. "input_cost_per_token": 0.000008,
  480. "output_cost_per_token": 0.000024
  481. },
  482. "anthropic/claude-2.1": {
  483. "max_tokens": 8191,
  484. "max_input_tokens": 200000,
  485. "max_output_tokens": 8191,
  486. "input_cost_per_token": 0.000008,
  487. "output_cost_per_token": 0.000024
  488. },
  489. "anthropic/claude-3-haiku-20240307": {
  490. "max_tokens": 4096,
  491. "max_input_tokens": 200000,
  492. "max_output_tokens": 4096,
  493. "input_cost_per_token": 0.00000025,
  494. "output_cost_per_token": 0.00000125
  495. },
  496. "anthropic/claude-3-opus-20240229": {
  497. "max_tokens": 4096,
  498. "max_input_tokens": 200000,
  499. "max_output_tokens": 4096,
  500. "input_cost_per_token": 0.000015,
  501. "output_cost_per_token": 0.000075
  502. },
  503. "anthropic/claude-3-sonnet-20240229": {
  504. "max_tokens": 4096,
  505. "max_input_tokens": 200000,
  506. "max_output_tokens": 4096,
  507. "input_cost_per_token": 0.000003,
  508. "output_cost_per_token": 0.000015
  509. },
  510. "vertexai/chat-bison": {
  511. "max_tokens": 4096,
  512. "max_input_tokens": 8192,
  513. "max_output_tokens": 4096,
  514. "input_cost_per_token": 0.000000125,
  515. "output_cost_per_token": 0.000000125
  516. },
  517. "vertexai/chat-bison@001": {
  518. "max_tokens": 4096,
  519. "max_input_tokens": 8192,
  520. "max_output_tokens": 4096,
  521. "input_cost_per_token": 0.000000125,
  522. "output_cost_per_token": 0.000000125
  523. },
  524. "vertexai/chat-bison@002": {
  525. "max_tokens": 4096,
  526. "max_input_tokens": 8192,
  527. "max_output_tokens": 4096,
  528. "input_cost_per_token": 0.000000125,
  529. "output_cost_per_token": 0.000000125
  530. },
  531. "vertexai/chat-bison-32k": {
  532. "max_tokens": 8192,
  533. "max_input_tokens": 32000,
  534. "max_output_tokens": 8192,
  535. "input_cost_per_token": 0.000000125,
  536. "output_cost_per_token": 0.000000125
  537. },
  538. "vertexai/code-bison": {
  539. "max_tokens": 1024,
  540. "max_input_tokens": 6144,
  541. "max_output_tokens": 1024,
  542. "input_cost_per_token": 0.000000125,
  543. "output_cost_per_token": 0.000000125
  544. },
  545. "vertexai/code-bison@001": {
  546. "max_tokens": 1024,
  547. "max_input_tokens": 6144,
  548. "max_output_tokens": 1024,
  549. "input_cost_per_token": 0.000000125,
  550. "output_cost_per_token": 0.000000125
  551. },
  552. "vertexai/code-gecko@001": {
  553. "max_tokens": 64,
  554. "max_input_tokens": 2048,
  555. "max_output_tokens": 64,
  556. "input_cost_per_token": 0.000000125,
  557. "output_cost_per_token": 0.000000125
  558. },
  559. "vertexai/code-gecko@002": {
  560. "max_tokens": 64,
  561. "max_input_tokens": 2048,
  562. "max_output_tokens": 64,
  563. "input_cost_per_token": 0.000000125,
  564. "output_cost_per_token": 0.000000125
  565. },
  566. "vertexai/code-gecko": {
  567. "max_tokens": 64,
  568. "max_input_tokens": 2048,
  569. "max_output_tokens": 64,
  570. "input_cost_per_token": 0.000000125,
  571. "output_cost_per_token": 0.000000125
  572. },
  573. "vertexai/codechat-bison": {
  574. "max_tokens": 1024,
  575. "max_input_tokens": 6144,
  576. "max_output_tokens": 1024,
  577. "input_cost_per_token": 0.000000125,
  578. "output_cost_per_token": 0.000000125
  579. },
  580. "vertexai/codechat-bison@001": {
  581. "max_tokens": 1024,
  582. "max_input_tokens": 6144,
  583. "max_output_tokens": 1024,
  584. "input_cost_per_token": 0.000000125,
  585. "output_cost_per_token": 0.000000125
  586. },
  587. "vertexai/codechat-bison-32k": {
  588. "max_tokens": 8192,
  589. "max_input_tokens": 32000,
  590. "max_output_tokens": 8192,
  591. "input_cost_per_token": 0.000000125,
  592. "output_cost_per_token": 0.000000125
  593. },
  594. "vertexai/gemini-pro": {
  595. "max_tokens": 8192,
  596. "max_input_tokens": 32760,
  597. "max_output_tokens": 8192,
  598. "input_cost_per_token": 0.00000025,
  599. "output_cost_per_token": 0.0000005
  600. },
  601. "vertexai/gemini-1.0-pro": {
  602. "max_tokens": 8192,
  603. "max_input_tokens": 32760,
  604. "max_output_tokens": 8192,
  605. "input_cost_per_token": 0.00000025,
  606. "output_cost_per_token": 0.0000005
  607. },
  608. "vertexai/gemini-1.0-pro-001": {
  609. "max_tokens": 8192,
  610. "max_input_tokens": 32760,
  611. "max_output_tokens": 8192,
  612. "input_cost_per_token": 0.00000025,
  613. "output_cost_per_token": 0.0000005
  614. },
  615. "vertexai/gemini-1.0-pro-002": {
  616. "max_tokens": 8192,
  617. "max_input_tokens": 32760,
  618. "max_output_tokens": 8192,
  619. "input_cost_per_token": 0.00000025,
  620. "output_cost_per_token": 0.0000005
  621. },
  622. "vertexai/gemini-1.5-pro": {
  623. "max_tokens": 8192,
  624. "max_input_tokens": 1000000,
  625. "max_output_tokens": 8192,
  626. "input_cost_per_token": 0.000000625,
  627. "output_cost_per_token": 0.000001875
  628. },
  629. "vertexai/gemini-1.5-flash-001": {
  630. "max_tokens": 8192,
  631. "max_input_tokens": 1000000,
  632. "max_output_tokens": 8192,
  633. "input_cost_per_token": 0,
  634. "output_cost_per_token": 0
  635. },
  636. "vertexai/gemini-1.5-flash-preview-0514": {
  637. "max_tokens": 8192,
  638. "max_input_tokens": 1000000,
  639. "max_output_tokens": 8192,
  640. "input_cost_per_token": 0,
  641. "output_cost_per_token": 0
  642. },
  643. "vertexai/gemini-1.5-pro-001": {
  644. "max_tokens": 8192,
  645. "max_input_tokens": 1000000,
  646. "max_output_tokens": 8192,
  647. "input_cost_per_token": 0.000000625,
  648. "output_cost_per_token": 0.000001875
  649. },
  650. "vertexai/gemini-1.5-pro-preview-0514": {
  651. "max_tokens": 8192,
  652. "max_input_tokens": 1000000,
  653. "max_output_tokens": 8192,
  654. "input_cost_per_token": 0.000000625,
  655. "output_cost_per_token": 0.000001875
  656. },
  657. "vertexai/gemini-1.5-pro-preview-0215": {
  658. "max_tokens": 8192,
  659. "max_input_tokens": 1000000,
  660. "max_output_tokens": 8192,
  661. "input_cost_per_token": 0.000000625,
  662. "output_cost_per_token": 0.000001875
  663. },
  664. "vertexai/gemini-1.5-pro-preview-0409": {
  665. "max_tokens": 8192,
  666. "max_input_tokens": 1000000,
  667. "max_output_tokens": 8192,
  668. "input_cost_per_token": 0.000000625,
  669. "output_cost_per_token": 0.000001875
  670. },
  671. "vertexai/gemini-experimental": {
  672. "max_tokens": 8192,
  673. "max_input_tokens": 1000000,
  674. "max_output_tokens": 8192,
  675. "input_cost_per_token": 0,
  676. "output_cost_per_token": 0
  677. },
  678. "vertexai/gemini-pro-vision": {
  679. "max_tokens": 2048,
  680. "max_input_tokens": 16384,
  681. "max_output_tokens": 2048,
  682. "max_images_per_prompt": 16,
  683. "max_videos_per_prompt": 1,
  684. "max_video_length": 2,
  685. "input_cost_per_token": 0.00000025,
  686. "output_cost_per_token": 0.0000005
  687. },
  688. "vertexai/gemini-1.0-pro-vision": {
  689. "max_tokens": 2048,
  690. "max_input_tokens": 16384,
  691. "max_output_tokens": 2048,
  692. "max_images_per_prompt": 16,
  693. "max_videos_per_prompt": 1,
  694. "max_video_length": 2,
  695. "input_cost_per_token": 0.00000025,
  696. "output_cost_per_token": 0.0000005
  697. },
  698. "vertexai/gemini-1.0-pro-vision-001": {
  699. "max_tokens": 2048,
  700. "max_input_tokens": 16384,
  701. "max_output_tokens": 2048,
  702. "max_images_per_prompt": 16,
  703. "max_videos_per_prompt": 1,
  704. "max_video_length": 2,
  705. "input_cost_per_token": 0.00000025,
  706. "output_cost_per_token": 0.0000005
  707. },
  708. "vertexai/claude-3-sonnet@20240229": {
  709. "max_tokens": 4096,
  710. "max_input_tokens": 200000,
  711. "max_output_tokens": 4096,
  712. "input_cost_per_token": 0.000003,
  713. "output_cost_per_token": 0.000015
  714. },
  715. "vertexai/claude-3-haiku@20240307": {
  716. "max_tokens": 4096,
  717. "max_input_tokens": 200000,
  718. "max_output_tokens": 4096,
  719. "input_cost_per_token": 0.00000025,
  720. "output_cost_per_token": 0.00000125
  721. },
  722. "vertexai/claude-3-opus@20240229": {
  723. "max_tokens": 4096,
  724. "max_input_tokens": 200000,
  725. "max_output_tokens": 4096,
  726. "input_cost_per_token": 0.000015,
  727. "output_cost_per_token": 0.000075
  728. },
  729. "cohere/command-r": {
  730. "max_tokens": 4096,
  731. "max_input_tokens": 128000,
  732. "max_output_tokens": 4096,
  733. "input_cost_per_token": 0.00000050,
  734. "output_cost_per_token": 0.0000015
  735. },
  736. "cohere/command-light": {
  737. "max_tokens": 4096,
  738. "max_input_tokens": 4096,
  739. "max_output_tokens": 4096,
  740. "input_cost_per_token": 0.000015,
  741. "output_cost_per_token": 0.000015
  742. },
  743. "cohere/command-r-plus": {
  744. "max_tokens": 4096,
  745. "max_input_tokens": 128000,
  746. "max_output_tokens": 4096,
  747. "input_cost_per_token": 0.000003,
  748. "output_cost_per_token": 0.000015
  749. },
  750. "cohere/command-nightly": {
  751. "max_tokens": 4096,
  752. "max_input_tokens": 4096,
  753. "max_output_tokens": 4096,
  754. "input_cost_per_token": 0.000015,
  755. "output_cost_per_token": 0.000015
  756. },
  757. "cohere/command": {
  758. "max_tokens": 4096,
  759. "max_input_tokens": 4096,
  760. "max_output_tokens": 4096,
  761. "input_cost_per_token": 0.000015,
  762. "output_cost_per_token": 0.000015
  763. },
  764. "cohere/command-medium-beta": {
  765. "max_tokens": 4096,
  766. "max_input_tokens": 4096,
  767. "max_output_tokens": 4096,
  768. "input_cost_per_token": 0.000015,
  769. "output_cost_per_token": 0.000015
  770. },
  771. "cohere/command-xlarge-beta": {
  772. "max_tokens": 4096,
  773. "max_input_tokens": 4096,
  774. "max_output_tokens": 4096,
  775. "input_cost_per_token": 0.000015,
  776. "output_cost_per_token": 0.000015
  777. },
  778. "together/together-ai-up-to-3b": {
  779. "input_cost_per_token": 0.0000001,
  780. "output_cost_per_token": 0.0000001
  781. },
  782. "together/together-ai-3.1b-7b": {
  783. "input_cost_per_token": 0.0000002,
  784. "output_cost_per_token": 0.0000002
  785. },
  786. "together/together-ai-7.1b-20b": {
  787. "max_tokens": 1000,
  788. "input_cost_per_token": 0.0000004,
  789. "output_cost_per_token": 0.0000004
  790. },
  791. "together/together-ai-20.1b-40b": {
  792. "input_cost_per_token": 0.0000008,
  793. "output_cost_per_token": 0.0000008
  794. },
  795. "together/together-ai-40.1b-70b": {
  796. "input_cost_per_token": 0.0000009,
  797. "output_cost_per_token": 0.0000009
  798. },
  799. "together/mistralai/Mixtral-8x7B-Instruct-v0.1": {
  800. "input_cost_per_token": 0.0000006,
  801. "output_cost_per_token": 0.0000006
  802. }
  803. }