[{"data":1,"prerenderedAt":844},["ShallowReactive",2],{"docs-/docs/fast-models":3},{"id":4,"title":5,"body":6,"description":837,"extension":838,"meta":839,"navigation":371,"path":840,"seo":841,"stem":842,"__hash__":843},"content/en/docs/fast-models.md","Fast Models - Seedance 2.0",{"type":7,"value":8,"toc":827},"minimark",[9,14,34,39,108,111,121,125,130,171,176,193,197,200,206,213,217,337,346,350,646,650,768,772,792,796,823],[10,11,13],"h1",{"id":12},"fast-models","Fast Models",[15,16,17,18,22,23,26,27,22,30,33],"p",{},"Every Seedance 2.0 mode ships in two tiers: ",[19,20,21],"strong",{},"Standard"," and ",[19,24,25],{},"Fast",". The Fast family trades a small amount of visual quality for ",[19,28,29],{},"faster generation",[19,31,32],{},"lower per-second pricing"," — perfect for rapid iteration, bulk production, and A/B testing.",[35,36,38],"h2",{"id":37},"the-three-fast-models","The Three Fast Models",[40,41,42,58],"table",{},[43,44,45],"thead",{},[46,47,48,52,55],"tr",{},[49,50,51],"th",{},"Fast Model ID",[49,53,54],{},"Corresponding Standard",[49,56,57],{},"Use for",[59,60,61,78,93],"tbody",{},[46,62,63,70,75],{},[64,65,66],"td",{},[67,68,69],"code",{},"seedance-2.0-fast-text-to-video",[64,71,72],{},[67,73,74],{},"seedance-2.0-text-to-video",[64,76,77],{},"Pure text-to-video",[46,79,80,85,90],{},[64,81,82],{},[67,83,84],{},"seedance-2.0-fast-image-to-video",[64,86,87],{},[67,88,89],{},"seedance-2.0-image-to-video",[64,91,92],{},"1- or 2-image driven",[46,94,95,100,105],{},[64,96,97],{},[67,98,99],{},"seedance-2.0-fast-reference-to-video",[64,101,102],{},[67,103,104],{},"seedance-2.0-reference-to-video",[64,106,107],{},"Multimodal composition",[15,109,110],{},"All Fast models share the same endpoint and parameter structure:",[112,113,118],"pre",{"className":114,"code":116,"language":117},[115],"language-text","POST https://api.evolink.ai/v1/videos/generations\n","text",[67,119,116],{"__ignoreMap":120},"",[35,122,124],{"id":123},"differences-from-standard","Differences from Standard",[15,126,127],{},[19,128,129],{},"Identical:",[131,132,133,137,140,159,162,165],"ul",{},[134,135,136],"li",{},"Endpoint",[134,138,139],{},"Request body schema (all parameter names, types, default values)",[134,141,142,143,146,147,150,151,154,155,158],{},"Allowed quality tiers (",[67,144,145],{},"480p"," / ",[67,148,149],{},"720p","), duration range (",[67,152,153],{},"4","–",[67,156,157],{},"15"," seconds), aspect ratios",[134,160,161],{},"Input asset count and format limits",[134,163,164],{},"Response schema, task lifecycle, webhook payload format",[134,166,167,170],{},[67,168,169],{},"prompt"," length limit (500 Chinese chars / 1000 English words)",[15,172,173],{},[19,174,175],{},"Different:",[131,177,178,181,184,187],{},[134,179,180],{},"Faster generation",[134,182,183],{},"Lower per-second pricing",[134,185,186],{},"Slightly lower quality detail than Standard (usually indistinguishable to the eye)",[134,188,189,192],{},[67,190,191],{},"fast-image-to-video"," auto-detects first-frame vs first-last-frame mode based on image count (1 = first-frame driven, 2 = first-last-frame transition); no extra field needed",[35,194,196],{"id":195},"recommended-workflow","Recommended Workflow",[15,198,199],{},"A typical production pipeline uses both tiers:",[112,201,204],{"className":202,"code":203,"language":117},[115],"Iterate on prompts / parameters\n    ↓  (Fast model — quick & cheap)\n    ↓\nIdentify the prompts and params you're happy with\n    ↓  (Swap `model` field, keep everything else)\n    ↓\nFinal delivery\n    ↓  (Standard model — highest quality render)\n",[67,205,203],{"__ignoreMap":120},[15,207,208,209,212],{},"Code-wise, the only change is the ",[67,210,211],{},"model"," string — no other logic needs to change.",[35,214,216],{"id":215},"example-fast-text-to-video-web_search","Example: Fast Text-to-Video + web_search",[112,218,222],{"className":219,"code":220,"language":221,"meta":120,"style":120},"language-bash shiki shiki-themes github-dark","curl -X POST https://api.evolink.ai/v1/videos/generations \\\n  -H \"Authorization: Bearer YOUR_API_KEY\" \\\n  -H \"Content-Type: application/json\" \\\n  -d '{\n    \"model\": \"seedance-2.0-fast-text-to-video\",\n    \"prompt\": \"A commercial introducing the latest 2026 electric sports car, highlighting its aerodynamic design and cabin tech.\",\n    \"duration\": 6,\n    \"quality\": \"720p\",\n    \"aspect_ratio\": \"16:9\",\n    \"generate_audio\": true,\n    \"model_params\": {\n      \"web_search\": true\n    }\n  }'\n","bash",[67,223,224,247,258,268,277,283,289,295,301,307,313,319,325,331],{"__ignoreMap":120},[225,226,229,233,237,241,244],"span",{"class":227,"line":228},"line",1,[225,230,232],{"class":231},"svObZ","curl",[225,234,236],{"class":235},"sDLfK"," -X",[225,238,240],{"class":239},"sU2Wk"," POST",[225,242,243],{"class":239}," https://api.evolink.ai/v1/videos/generations",[225,245,246],{"class":235}," \\\n",[225,248,250,253,256],{"class":227,"line":249},2,[225,251,252],{"class":235},"  -H",[225,254,255],{"class":239}," \"Authorization: Bearer YOUR_API_KEY\"",[225,257,246],{"class":235},[225,259,261,263,266],{"class":227,"line":260},3,[225,262,252],{"class":235},[225,264,265],{"class":239}," \"Content-Type: application/json\"",[225,267,246],{"class":235},[225,269,271,274],{"class":227,"line":270},4,[225,272,273],{"class":235},"  -d",[225,275,276],{"class":239}," '{\n",[225,278,280],{"class":227,"line":279},5,[225,281,282],{"class":239},"    \"model\": \"seedance-2.0-fast-text-to-video\",\n",[225,284,286],{"class":227,"line":285},6,[225,287,288],{"class":239},"    \"prompt\": \"A commercial introducing the latest 2026 electric sports car, highlighting its aerodynamic design and cabin tech.\",\n",[225,290,292],{"class":227,"line":291},7,[225,293,294],{"class":239},"    \"duration\": 6,\n",[225,296,298],{"class":227,"line":297},8,[225,299,300],{"class":239},"    \"quality\": \"720p\",\n",[225,302,304],{"class":227,"line":303},9,[225,305,306],{"class":239},"    \"aspect_ratio\": \"16:9\",\n",[225,308,310],{"class":227,"line":309},10,[225,311,312],{"class":239},"    \"generate_audio\": true,\n",[225,314,316],{"class":227,"line":315},11,[225,317,318],{"class":239},"    \"model_params\": {\n",[225,320,322],{"class":227,"line":321},12,[225,323,324],{"class":239},"      \"web_search\": true\n",[225,326,328],{"class":227,"line":327},13,[225,329,330],{"class":239},"    }\n",[225,332,334],{"class":227,"line":333},14,[225,335,336],{"class":239},"  }'\n",[338,339,340],"blockquote",{},[15,341,342,345],{},[67,343,344],{},"model_params.web_search"," is exclusive to the text-to-video family (including Fast). It's only billed when a search is actually triggered.",[35,347,349],{"id":348},"example-fast-image-to-video-auto-first-frame-first-last-frame","Example: Fast Image-to-Video (auto first-frame / first-last-frame)",[112,351,355],{"className":352,"code":353,"language":354,"meta":120,"style":120},"language-python shiki shiki-themes github-dark","import requests\n\n# 1 image → first-frame driven\nresponse = requests.post(\n    \"https://api.evolink.ai/v1/videos/generations\",\n    headers={\n        \"Authorization\": \"Bearer YOUR_API_KEY\",\n        \"Content-Type\": \"application/json\"\n    },\n    json={\n        \"model\": \"seedance-2.0-fast-image-to-video\",\n        \"prompt\": \"Camera slowly pushes in, the scene comes alive\",\n        \"image_urls\": [\"https://example.com/scene.jpg\"],\n        \"duration\": 5\n    }\n)\n\n# 2 images → auto switches to first-last-frame transition\nresponse = requests.post(\n    \"https://api.evolink.ai/v1/videos/generations\",\n    headers={\n        \"Authorization\": \"Bearer YOUR_API_KEY\",\n        \"Content-Type\": \"application/json\"\n    },\n    json={\n        \"model\": \"seedance-2.0-fast-image-to-video\",\n        \"prompt\": \"A smooth transition between two scenes\",\n        \"image_urls\": [\n            \"https://example.com/first.jpg\",\n            \"https://example.com/last.jpg\"\n        ],\n        \"duration\": 6\n    }\n)\n","python",[67,356,357,367,373,379,390,398,409,422,432,437,446,458,470,484,494,499,505,510,516,525,532,541,552,561,566,575,586,598,606,614,620,626,636,641],{"__ignoreMap":120},[225,358,359,363],{"class":227,"line":228},[225,360,362],{"class":361},"snl16","import",[225,364,366],{"class":365},"s95oV"," requests\n",[225,368,369],{"class":227,"line":249},[225,370,372],{"emptyLinePlaceholder":371},true,"\n",[225,374,375],{"class":227,"line":260},[225,376,378],{"class":377},"sAwPA","# 1 image → first-frame driven\n",[225,380,381,384,387],{"class":227,"line":270},[225,382,383],{"class":365},"response ",[225,385,386],{"class":361},"=",[225,388,389],{"class":365}," requests.post(\n",[225,391,392,395],{"class":227,"line":279},[225,393,394],{"class":239},"    \"https://api.evolink.ai/v1/videos/generations\"",[225,396,397],{"class":365},",\n",[225,399,400,404,406],{"class":227,"line":285},[225,401,403],{"class":402},"s9osk","    headers",[225,405,386],{"class":361},[225,407,408],{"class":365},"{\n",[225,410,411,414,417,420],{"class":227,"line":291},[225,412,413],{"class":239},"        \"Authorization\"",[225,415,416],{"class":365},": ",[225,418,419],{"class":239},"\"Bearer YOUR_API_KEY\"",[225,421,397],{"class":365},[225,423,424,427,429],{"class":227,"line":297},[225,425,426],{"class":239},"        \"Content-Type\"",[225,428,416],{"class":365},[225,430,431],{"class":239},"\"application/json\"\n",[225,433,434],{"class":227,"line":303},[225,435,436],{"class":365},"    },\n",[225,438,439,442,444],{"class":227,"line":309},[225,440,441],{"class":402},"    json",[225,443,386],{"class":361},[225,445,408],{"class":365},[225,447,448,451,453,456],{"class":227,"line":315},[225,449,450],{"class":239},"        \"model\"",[225,452,416],{"class":365},[225,454,455],{"class":239},"\"seedance-2.0-fast-image-to-video\"",[225,457,397],{"class":365},[225,459,460,463,465,468],{"class":227,"line":321},[225,461,462],{"class":239},"        \"prompt\"",[225,464,416],{"class":365},[225,466,467],{"class":239},"\"Camera slowly pushes in, the scene comes alive\"",[225,469,397],{"class":365},[225,471,472,475,478,481],{"class":227,"line":327},[225,473,474],{"class":239},"        \"image_urls\"",[225,476,477],{"class":365},": [",[225,479,480],{"class":239},"\"https://example.com/scene.jpg\"",[225,482,483],{"class":365},"],\n",[225,485,486,489,491],{"class":227,"line":333},[225,487,488],{"class":239},"        \"duration\"",[225,490,416],{"class":365},[225,492,493],{"class":235},"5\n",[225,495,497],{"class":227,"line":496},15,[225,498,330],{"class":365},[225,500,502],{"class":227,"line":501},16,[225,503,504],{"class":365},")\n",[225,506,508],{"class":227,"line":507},17,[225,509,372],{"emptyLinePlaceholder":371},[225,511,513],{"class":227,"line":512},18,[225,514,515],{"class":377},"# 2 images → auto switches to first-last-frame transition\n",[225,517,519,521,523],{"class":227,"line":518},19,[225,520,383],{"class":365},[225,522,386],{"class":361},[225,524,389],{"class":365},[225,526,528,530],{"class":227,"line":527},20,[225,529,394],{"class":239},[225,531,397],{"class":365},[225,533,535,537,539],{"class":227,"line":534},21,[225,536,403],{"class":402},[225,538,386],{"class":361},[225,540,408],{"class":365},[225,542,544,546,548,550],{"class":227,"line":543},22,[225,545,413],{"class":239},[225,547,416],{"class":365},[225,549,419],{"class":239},[225,551,397],{"class":365},[225,553,555,557,559],{"class":227,"line":554},23,[225,556,426],{"class":239},[225,558,416],{"class":365},[225,560,431],{"class":239},[225,562,564],{"class":227,"line":563},24,[225,565,436],{"class":365},[225,567,569,571,573],{"class":227,"line":568},25,[225,570,441],{"class":402},[225,572,386],{"class":361},[225,574,408],{"class":365},[225,576,578,580,582,584],{"class":227,"line":577},26,[225,579,450],{"class":239},[225,581,416],{"class":365},[225,583,455],{"class":239},[225,585,397],{"class":365},[225,587,589,591,593,596],{"class":227,"line":588},27,[225,590,462],{"class":239},[225,592,416],{"class":365},[225,594,595],{"class":239},"\"A smooth transition between two scenes\"",[225,597,397],{"class":365},[225,599,601,603],{"class":227,"line":600},28,[225,602,474],{"class":239},[225,604,605],{"class":365},": [\n",[225,607,609,612],{"class":227,"line":608},29,[225,610,611],{"class":239},"            \"https://example.com/first.jpg\"",[225,613,397],{"class":365},[225,615,617],{"class":227,"line":616},30,[225,618,619],{"class":239},"            \"https://example.com/last.jpg\"\n",[225,621,623],{"class":227,"line":622},31,[225,624,625],{"class":365},"        ],\n",[225,627,629,631,633],{"class":227,"line":628},32,[225,630,488],{"class":239},[225,632,416],{"class":365},[225,634,635],{"class":235},"6\n",[225,637,639],{"class":227,"line":638},33,[225,640,330],{"class":365},[225,642,644],{"class":227,"line":643},34,[225,645,504],{"class":365},[35,647,649],{"id":648},"example-fast-reference-to-video","Example: Fast Reference-to-Video",[112,651,653],{"className":219,"code":652,"language":221,"meta":120,"style":120},"curl -X POST https://api.evolink.ai/v1/videos/generations \\\n  -H \"Authorization: Bearer YOUR_API_KEY\" \\\n  -H \"Content-Type: application/json\" \\\n  -d '{\n    \"model\": \"seedance-2.0-fast-reference-to-video\",\n    \"prompt\": \"Replicate video 1's first-person perspective. Use audio 1 as background music throughout. Promo video opening.\",\n    \"image_urls\": [\"https://example.com/ref1.jpg\"],\n    \"video_urls\": [\"https://example.com/reference.mp4\"],\n    \"audio_urls\": [\"https://example.com/bgm.mp3\"],\n    \"duration\": 10,\n    \"quality\": \"720p\",\n    \"aspect_ratio\": \"16:9\"\n  }'\n",[67,654,655,667,675,683,689,694,735,740,745,750,755,759,764],{"__ignoreMap":120},[225,656,657,659,661,663,665],{"class":227,"line":228},[225,658,232],{"class":231},[225,660,236],{"class":235},[225,662,240],{"class":239},[225,664,243],{"class":239},[225,666,246],{"class":235},[225,668,669,671,673],{"class":227,"line":249},[225,670,252],{"class":235},[225,672,255],{"class":239},[225,674,246],{"class":235},[225,676,677,679,681],{"class":227,"line":260},[225,678,252],{"class":235},[225,680,265],{"class":239},[225,682,246],{"class":235},[225,684,685,687],{"class":227,"line":270},[225,686,273],{"class":235},[225,688,276],{"class":239},[225,690,691],{"class":227,"line":279},[225,692,693],{"class":239},"    \"model\": \"seedance-2.0-fast-reference-to-video\",\n",[225,695,696,699,702,705,708,711,714,717,720,723,726,729,732],{"class":227,"line":285},[225,697,698],{"class":239},"    \"prompt\": \"Replicate video 1's",[225,700,701],{"class":239}," first-person",[225,703,704],{"class":239}," perspective.",[225,706,707],{"class":239}," Use",[225,709,710],{"class":239}," audio",[225,712,713],{"class":235}," 1",[225,715,716],{"class":239}," as",[225,718,719],{"class":239}," background",[225,721,722],{"class":239}," music",[225,724,725],{"class":239}," throughout.",[225,727,728],{"class":239}," Promo",[225,730,731],{"class":239}," video",[225,733,734],{"class":239}," opening.\",\n",[225,736,737],{"class":227,"line":291},[225,738,739],{"class":239},"    \"image_urls\": [\"https://example.com/ref1.jpg\"],\n",[225,741,742],{"class":227,"line":297},[225,743,744],{"class":239},"    \"video_urls\": [\"https://example.com/reference.mp4\"],\n",[225,746,747],{"class":227,"line":303},[225,748,749],{"class":239},"    \"audio_urls\": [\"https://example.com/bgm.mp3\"],\n",[225,751,752],{"class":227,"line":309},[225,753,754],{"class":239},"    \"duration\": 10,\n",[225,756,757],{"class":227,"line":315},[225,758,300],{"class":239},[225,760,761],{"class":227,"line":321},[225,762,763],{"class":239},"    \"aspect_ratio\": \"16:9\"\n",[225,765,766],{"class":227,"line":327},[225,767,336],{"class":239},[35,769,771],{"id":770},"when-not-to-use-fast","When Not to Use Fast",[131,773,774,780,786],{},[134,775,776,779],{},[19,777,778],{},"Final ad deliverables / brand hero films"," — choose Standard for more stable detail",[134,781,782,785],{},[19,783,784],{},"Close-up face or micro-expression shots"," — Standard is more precise",[134,787,788,791],{},[19,789,790],{},"Complex reference-to-video composition"," with 9 images + 3 videos + 3 audio clips — Standard comprehends the combined signals better",[35,793,795],{"id":794},"related","Related",[131,797,798,805,811,817],{},[134,799,800],{},[801,802,804],"a",{"href":803},"/docs/models-overview","Models Overview",[134,806,807],{},[801,808,810],{"href":809},"/docs/text-to-video","Text-to-Video API",[134,812,813],{},[801,814,816],{"href":815},"/docs/image-to-video","Image-to-Video API",[134,818,819],{},[801,820,822],{"href":821},"/docs/reference-to-video","Reference-to-Video API",[824,825,826],"style",{},"html pre.shiki code .svObZ, html code.shiki .svObZ{--shiki-default:#B392F0}html pre.shiki code .sDLfK, html code.shiki .sDLfK{--shiki-default:#79B8FF}html pre.shiki code .sU2Wk, html code.shiki .sU2Wk{--shiki-default:#9ECBFF}html .default .shiki span {color: var(--shiki-default);background: var(--shiki-default-bg);font-style: var(--shiki-default-font-style);font-weight: var(--shiki-default-font-weight);text-decoration: var(--shiki-default-text-decoration);}html .shiki span {color: var(--shiki-default);background: var(--shiki-default-bg);font-style: var(--shiki-default-font-style);font-weight: var(--shiki-default-font-weight);text-decoration: var(--shiki-default-text-decoration);}html pre.shiki code .snl16, html code.shiki .snl16{--shiki-default:#F97583}html pre.shiki code .s95oV, html code.shiki .s95oV{--shiki-default:#E1E4E8}html pre.shiki code .sAwPA, html code.shiki .sAwPA{--shiki-default:#6A737D}html pre.shiki code .s9osk, html code.shiki .s9osk{--shiki-default:#FFAB70}",{"title":120,"searchDepth":249,"depth":249,"links":828},[829,830,831,832,833,834,835,836],{"id":37,"depth":249,"text":38},{"id":123,"depth":249,"text":124},{"id":195,"depth":249,"text":196},{"id":215,"depth":249,"text":216},{"id":348,"depth":249,"text":349},{"id":648,"depth":249,"text":649},{"id":770,"depth":249,"text":771},{"id":794,"depth":249,"text":795},"Seedance 2.0 Fast model family (fast-text-to-video / fast-image-to-video / fast-reference-to-video): faster generation, lower per-second cost, identical API shape.","md",{},"/en/docs/fast-models",{"title":5,"description":837},"en/docs/fast-models","tcpExlr3fSOVncixjStmNWKmZWzaOMMkJZqqrGCUxuo",1776086321903]