[{"data":1,"prerenderedAt":538},["ShallowReactive",2],{"model-/models/seedance-2-0-text-to-video":3},{"id":4,"title":5,"body":6,"description":513,"extension":514,"meta":515,"navigation":201,"path":534,"seo":535,"stem":536,"__hash__":537},"content/en/models/seedance-2-0-text-to-video.md","Seedance 2.0 Text-to-Video — Generate AI Video from Text Prompts",{"type":7,"value":8,"toc":498},"minimark",[9,14,18,38,42,74,78,83,86,90,145,149,154,160,164,170,174,358,362,368,428,436,440,466,470,494],[10,11,13],"h2",{"id":12},"what-is-text-to-video","What Is Text-to-Video?",[15,16,17],"p",{},"Seedance 2.0's text-to-video capability turns natural language descriptions into high-quality cinematic video. Describe the scene, motion, and style you want, and the model generates a matching video.",[15,19,20,24,25,29,30,37],{},[21,22,23],"strong",{},"Model ID:"," ",[26,27,28],"code",{},"seedance-2.0-text-to-video"," (for faster generation and lower cost, use ",[31,32,34],"a",{"href":33},"/docs/fast-models",[26,35,36],{},"seedance-2.0-fast-text-to-video",")",[10,39,41],{"id":40},"how-it-works","How It Works",[43,44,45,52,58,68],"ol",{},[46,47,48,51],"li",{},[21,49,50],{},"Write the prompt"," — describe the scene you want",[46,53,54,57],{},[21,55,56],{},"Send the API request"," — model, prompt, quality, duration, and other parameters",[46,59,60,63,64,67],{},[21,61,62],{},"Get a task ID"," — poll asynchronously or configure a ",[26,65,66],{},"callback_url"," webhook",[46,69,70,73],{},[21,71,72],{},"Download the video"," — retrieve the URL when generation completes (valid 24 hours)",[10,75,77],{"id":76},"prompt-engineering-tips","Prompt Engineering Tips",[79,80,82],"h3",{"id":81},"be-specific","Be Specific",[15,84,85],{},"Don't just say \"a dog running\". Write: \"A golden retriever running through a sunlit meadow, slow motion, shallow depth of field, cinematic.\"",[79,87,89],{"id":88},"use-style-keywords","Use Style Keywords",[91,92,93,111,128],"ul",{},[46,94,95,24,98,101,102,101,105,101,108],{},[21,96,97],{},"Shots:",[26,99,100],{},"aerial shot",", ",[26,103,104],{},"close-up",[26,106,107],{},"tracking shot",[26,109,110],{},"dolly zoom",[46,112,113,24,116,101,119,101,122,101,125],{},[21,114,115],{},"Lighting:",[26,117,118],{},"golden hour",[26,120,121],{},"studio lighting",[26,123,124],{},"neon",[26,126,127],{},"dramatic shadows",[46,129,130,24,133,101,136,101,139,101,142],{},[21,131,132],{},"Style:",[26,134,135],{},"cinematic",[26,137,138],{},"slow motion",[26,140,141],{},"timelapse",[26,143,144],{},"documentary",[79,146,148],{"id":147},"structure-your-prompt","Structure Your Prompt",[15,150,151],{},[26,152,153],{},"[Subject] + [Action] + [Setting] + [Style/Mood]",[15,155,156,157],{},"Example: ",[26,158,159],{},"A woman in a red dress walking through a rain-soaked Tokyo street at night, neon reflections on wet pavement, cinematic wide shot",[79,161,163],{"id":162},"put-dialogue-in-double-quotes","Put Dialogue in Double Quotes",[15,165,166,167],{},"Wrap spoken lines in straight double quotes. The model detects them as dialogue and optimizes speech synthesis:\n",[26,168,169],{},"She turned and said: \"You finally understood.\"",[10,171,173],{"id":172},"api-reference","API Reference",[175,176,181],"pre",{"className":177,"code":178,"language":179,"meta":180,"style":180},"language-python shiki shiki-themes github-dark","import requests\n\nresponse = requests.post(\n    \"https://api.evolink.ai/v1/videos/generations\",\n    headers={\"Authorization\": \"Bearer YOUR_API_KEY\"},\n    json={\n        \"model\": \"seedance-2.0-text-to-video\",\n        \"prompt\": \"Your scene description here\",\n        \"quality\": \"720p\",       # 480p or 720p\n        \"duration\": 5,           # 4–15 seconds (integer)\n        \"aspect_ratio\": \"16:9\",\n        \"generate_audio\": True\n    }\n)\n","python","",[26,182,183,196,203,215,225,249,260,273,286,304,322,335,346,352],{"__ignoreMap":180},[184,185,188,192],"span",{"class":186,"line":187},"line",1,[184,189,191],{"class":190},"snl16","import",[184,193,195],{"class":194},"s95oV"," requests\n",[184,197,199],{"class":186,"line":198},2,[184,200,202],{"emptyLinePlaceholder":201},true,"\n",[184,204,206,209,212],{"class":186,"line":205},3,[184,207,208],{"class":194},"response ",[184,210,211],{"class":190},"=",[184,213,214],{"class":194}," requests.post(\n",[184,216,218,222],{"class":186,"line":217},4,[184,219,221],{"class":220},"sU2Wk","    \"https://api.evolink.ai/v1/videos/generations\"",[184,223,224],{"class":194},",\n",[184,226,228,232,234,237,240,243,246],{"class":186,"line":227},5,[184,229,231],{"class":230},"s9osk","    headers",[184,233,211],{"class":190},[184,235,236],{"class":194},"{",[184,238,239],{"class":220},"\"Authorization\"",[184,241,242],{"class":194},": ",[184,244,245],{"class":220},"\"Bearer YOUR_API_KEY\"",[184,247,248],{"class":194},"},\n",[184,250,252,255,257],{"class":186,"line":251},6,[184,253,254],{"class":230},"    json",[184,256,211],{"class":190},[184,258,259],{"class":194},"{\n",[184,261,263,266,268,271],{"class":186,"line":262},7,[184,264,265],{"class":220},"        \"model\"",[184,267,242],{"class":194},[184,269,270],{"class":220},"\"seedance-2.0-text-to-video\"",[184,272,224],{"class":194},[184,274,276,279,281,284],{"class":186,"line":275},8,[184,277,278],{"class":220},"        \"prompt\"",[184,280,242],{"class":194},[184,282,283],{"class":220},"\"Your scene description here\"",[184,285,224],{"class":194},[184,287,289,292,294,297,300],{"class":186,"line":288},9,[184,290,291],{"class":220},"        \"quality\"",[184,293,242],{"class":194},[184,295,296],{"class":220},"\"720p\"",[184,298,299],{"class":194},",       ",[184,301,303],{"class":302},"sAwPA","# 480p or 720p\n",[184,305,307,310,312,316,319],{"class":186,"line":306},10,[184,308,309],{"class":220},"        \"duration\"",[184,311,242],{"class":194},[184,313,315],{"class":314},"sDLfK","5",[184,317,318],{"class":194},",           ",[184,320,321],{"class":302},"# 4–15 seconds (integer)\n",[184,323,325,328,330,333],{"class":186,"line":324},11,[184,326,327],{"class":220},"        \"aspect_ratio\"",[184,329,242],{"class":194},[184,331,332],{"class":220},"\"16:9\"",[184,334,224],{"class":194},[184,336,338,341,343],{"class":186,"line":337},12,[184,339,340],{"class":220},"        \"generate_audio\"",[184,342,242],{"class":194},[184,344,345],{"class":314},"True\n",[184,347,349],{"class":186,"line":348},13,[184,350,351],{"class":194},"    }\n",[184,353,355],{"class":186,"line":354},14,[184,356,357],{"class":194},")\n",[79,359,361],{"id":360},"web-search-optional","Web Search (Optional)",[15,363,364,367],{},[26,365,366],{},"model_params.web_search"," is exclusive to the text-to-video family. When enabled, the model autonomously decides whether to search the internet for fresh content and is only billed when a search is actually triggered:",[175,369,371],{"className":177,"code":370,"language":179,"meta":180,"style":180},"json={\n    \"model\": \"seedance-2.0-text-to-video\",\n    \"prompt\": \"An ad for the latest 2026 electric sports car\",\n    \"model_params\": {\"web_search\": True}\n}\n",[26,372,373,382,393,405,424],{"__ignoreMap":180},[184,374,375,378,380],{"class":186,"line":187},[184,376,377],{"class":194},"json",[184,379,211],{"class":190},[184,381,259],{"class":194},[184,383,384,387,389,391],{"class":186,"line":198},[184,385,386],{"class":220},"    \"model\"",[184,388,242],{"class":194},[184,390,270],{"class":220},[184,392,224],{"class":194},[184,394,395,398,400,403],{"class":186,"line":205},[184,396,397],{"class":220},"    \"prompt\"",[184,399,242],{"class":194},[184,401,402],{"class":220},"\"An ad for the latest 2026 electric sports car\"",[184,404,224],{"class":194},[184,406,407,410,413,416,418,421],{"class":186,"line":217},[184,408,409],{"class":220},"    \"model_params\"",[184,411,412],{"class":194},": {",[184,414,415],{"class":220},"\"web_search\"",[184,417,242],{"class":194},[184,419,420],{"class":314},"True",[184,422,423],{"class":194},"}\n",[184,425,426],{"class":186,"line":227},[184,427,423],{"class":194},[15,429,430,431,435],{},"See the full parameter reference in the ",[31,432,434],{"href":433},"/docs/text-to-video","Text-to-Video API docs",".",[10,437,439],{"id":438},"use-cases","Use Cases",[91,441,442,448,454,460],{},[46,443,444,447],{},[21,445,446],{},"E-commerce:"," Generate product showcase videos from descriptions",[46,449,450,453],{},[21,451,452],{},"Marketing:"," Create ad creatives and social media content at scale",[46,455,456,459],{},[21,457,458],{},"Entertainment:"," Rapid prototyping of scenes and storyboards",[46,461,462,465],{},[21,463,464],{},"Education:"," Illustrate abstract concepts with custom videos",[10,467,469],{"id":468},"related","Related",[91,471,472,478,484,489],{},[46,473,474],{},[31,475,477],{"href":476},"/models/seedance-2-0","Seedance 2.0 Overview",[46,479,480],{},[31,481,483],{"href":482},"/models/seedance-2-0-image-to-video","Image-to-Video",[46,485,486],{},[31,487,488],{"href":433},"Text-to-Video API full reference",[46,490,491],{},[31,492,493],{"href":33},"Fast Models",[495,496,497],"style",{},"html pre.shiki code .snl16, html code.shiki .snl16{--shiki-default:#F97583}html pre.shiki code .s95oV, html code.shiki .s95oV{--shiki-default:#E1E4E8}html pre.shiki code .sU2Wk, html code.shiki .sU2Wk{--shiki-default:#9ECBFF}html pre.shiki code .s9osk, html code.shiki .s9osk{--shiki-default:#FFAB70}html pre.shiki code .sAwPA, html code.shiki .sAwPA{--shiki-default:#6A737D}html pre.shiki code .sDLfK, html code.shiki .sDLfK{--shiki-default:#79B8FF}html .default .shiki span {color: var(--shiki-default);background: var(--shiki-default-bg);font-style: var(--shiki-default-font-style);font-weight: var(--shiki-default-font-weight);text-decoration: var(--shiki-default-text-decoration);}html .shiki span {color: var(--shiki-default);background: var(--shiki-default-bg);font-style: var(--shiki-default-font-style);font-weight: var(--shiki-default-font-weight);text-decoration: var(--shiki-default-text-decoration);}",{"title":180,"searchDepth":198,"depth":198,"links":499},[500,501,502,508,511,512],{"id":12,"depth":198,"text":13},{"id":40,"depth":198,"text":41},{"id":76,"depth":198,"text":77,"children":503},[504,505,506,507],{"id":81,"depth":205,"text":82},{"id":88,"depth":205,"text":89},{"id":147,"depth":205,"text":148},{"id":162,"depth":205,"text":163},{"id":172,"depth":198,"text":173,"children":509},[510],{"id":360,"depth":205,"text":361},{"id":438,"depth":198,"text":439},{"id":468,"depth":198,"text":469},"Generate cinematic video from natural language descriptions using seedance-2.0-text-to-video. Supports 480p/720p, 4–15 second durations, and optional web_search augmentation.","md",{"modelName":516,"tags":517,"metrics":521},"Seedance 2.0 Text-to-Video",[518,519,520],"Text-to-Video","Seedance 2.0","Cinematic",[522,525,528,531],{"value":523,"label":524},"720p","Max Resolution",{"value":526,"label":527},"1000","Max English Words",{"value":529,"label":530},"15s","Max Duration",{"value":532,"label":533},"30-60s","Generation Time","/en/models/seedance-2-0-text-to-video",{"title":5,"description":513},"en/models/seedance-2-0-text-to-video","DfxEEmxc1bauFpKpdX2mbZIXZwkeA1Fp67ej-r72hmo",1776086320038]