[{"data":1,"prerenderedAt":537},["ShallowReactive",2],{"model-/es/models/seedance-2-0-text-to-video":3},{"id":4,"title":5,"body":6,"description":512,"extension":513,"meta":514,"navigation":200,"path":533,"seo":534,"stem":535,"__hash__":536},"content/es/models/seedance-2-0-text-to-video.md","Seedance 2.0 Text-to-Video — Genera vídeo con IA desde prompts de texto",{"type":7,"value":8,"toc":497},"minimark",[9,14,18,38,42,73,77,82,85,89,144,148,153,159,163,169,173,357,361,367,427,435,439,465,469,493],[10,11,13],"h2",{"id":12},"qué-es-text-to-video","¿Qué es Text-to-Video?",[15,16,17],"p",{},"La capacidad text-to-video de Seedance 2.0 convierte descripciones en lenguaje natural en vídeo cinematográfico de alta calidad. Describe la escena, el movimiento y el estilo que quieres y el modelo generará un vídeo que los plasme.",[15,19,20,24,25,29,30,37],{},[21,22,23],"strong",{},"ID del modelo:"," ",[26,27,28],"code",{},"seedance-2.0-text-to-video"," (para una generación más rápida y un coste menor, usa ",[31,32,34],"a",{"href":33},"/docs/fast-models",[26,35,36],{},"seedance-2.0-fast-text-to-video",")",[10,39,41],{"id":40},"cómo-funciona","Cómo funciona",[43,44,45,52,58,67],"ol",{},[46,47,48,51],"li",{},[21,49,50],{},"Escribe el prompt"," — describe la escena que quieres",[46,53,54,57],{},[21,55,56],{},"Envía la solicitud a la API"," — modelo, prompt, calidad, duración y otros parámetros",[46,59,60,63,64],{},[21,61,62],{},"Obtén un ID de tarea"," — haz polling asíncrono o configura un webhook ",[26,65,66],{},"callback_url",[46,68,69,72],{},[21,70,71],{},"Descarga el vídeo"," — recupera la URL cuando termine la generación (válida 24 horas)",[10,74,76],{"id":75},"consejos-de-prompt-engineering","Consejos de prompt engineering",[78,79,81],"h3",{"id":80},"sé-específico","Sé específico",[15,83,84],{},"No digas simplemente \"un perro corriendo\". Escribe: \"Un golden retriever corriendo por un prado iluminado por el sol, cámara lenta, profundidad de campo reducida, cinematográfico\".",[78,86,88],{"id":87},"usa-palabras-clave-de-estilo","Usa palabras clave de estilo",[90,91,92,110,127],"ul",{},[46,93,94,24,97,100,101,100,104,100,107],{},[21,95,96],{},"Planos:",[26,98,99],{},"aerial shot",", ",[26,102,103],{},"close-up",[26,105,106],{},"tracking shot",[26,108,109],{},"dolly zoom",[46,111,112,24,115,100,118,100,121,100,124],{},[21,113,114],{},"Iluminación:",[26,116,117],{},"golden hour",[26,119,120],{},"studio lighting",[26,122,123],{},"neon",[26,125,126],{},"dramatic shadows",[46,128,129,24,132,100,135,100,138,100,141],{},[21,130,131],{},"Estilo:",[26,133,134],{},"cinematic",[26,136,137],{},"slow motion",[26,139,140],{},"timelapse",[26,142,143],{},"documentary",[78,145,147],{"id":146},"estructura-tu-prompt","Estructura tu prompt",[15,149,150],{},[26,151,152],{},"[Sujeto] + [Acción] + [Escenario] + [Estilo/Atmósfera]",[15,154,155,156],{},"Ejemplo: ",[26,157,158],{},"A woman in a red dress walking through a rain-soaked Tokyo street at night, neon reflections on wet pavement, cinematic wide shot",[78,160,162],{"id":161},"coloca-los-diálogos-entre-comillas-dobles","Coloca los diálogos entre comillas dobles",[15,164,165,166],{},"Envuelve las líneas habladas en comillas dobles rectas. El modelo las detecta como diálogo y optimiza la síntesis de voz:\n",[26,167,168],{},"She turned and said: \"You finally understood.\"",[10,170,172],{"id":171},"referencia-de-la-api","Referencia de la API",[174,175,180],"pre",{"className":176,"code":177,"language":178,"meta":179,"style":179},"language-python shiki shiki-themes github-dark","import requests\n\nresponse = requests.post(\n    \"https://api.evolink.ai/v1/videos/generations\",\n    headers={\"Authorization\": \"Bearer YOUR_API_KEY\"},\n    json={\n        \"model\": \"seedance-2.0-text-to-video\",\n        \"prompt\": \"Your scene description here\",\n        \"quality\": \"720p\",       # 480p o 720p\n        \"duration\": 5,           # 4–15 segundos (entero)\n        \"aspect_ratio\": \"16:9\",\n        \"generate_audio\": True\n    }\n)\n","python","",[26,181,182,195,202,214,224,248,259,272,285,303,321,334,345,351],{"__ignoreMap":179},[183,184,187,191],"span",{"class":185,"line":186},"line",1,[183,188,190],{"class":189},"snl16","import",[183,192,194],{"class":193},"s95oV"," requests\n",[183,196,198],{"class":185,"line":197},2,[183,199,201],{"emptyLinePlaceholder":200},true,"\n",[183,203,205,208,211],{"class":185,"line":204},3,[183,206,207],{"class":193},"response ",[183,209,210],{"class":189},"=",[183,212,213],{"class":193}," requests.post(\n",[183,215,217,221],{"class":185,"line":216},4,[183,218,220],{"class":219},"sU2Wk","    \"https://api.evolink.ai/v1/videos/generations\"",[183,222,223],{"class":193},",\n",[183,225,227,231,233,236,239,242,245],{"class":185,"line":226},5,[183,228,230],{"class":229},"s9osk","    headers",[183,232,210],{"class":189},[183,234,235],{"class":193},"{",[183,237,238],{"class":219},"\"Authorization\"",[183,240,241],{"class":193},": ",[183,243,244],{"class":219},"\"Bearer YOUR_API_KEY\"",[183,246,247],{"class":193},"},\n",[183,249,251,254,256],{"class":185,"line":250},6,[183,252,253],{"class":229},"    json",[183,255,210],{"class":189},[183,257,258],{"class":193},"{\n",[183,260,262,265,267,270],{"class":185,"line":261},7,[183,263,264],{"class":219},"        \"model\"",[183,266,241],{"class":193},[183,268,269],{"class":219},"\"seedance-2.0-text-to-video\"",[183,271,223],{"class":193},[183,273,275,278,280,283],{"class":185,"line":274},8,[183,276,277],{"class":219},"        \"prompt\"",[183,279,241],{"class":193},[183,281,282],{"class":219},"\"Your scene description here\"",[183,284,223],{"class":193},[183,286,288,291,293,296,299],{"class":185,"line":287},9,[183,289,290],{"class":219},"        \"quality\"",[183,292,241],{"class":193},[183,294,295],{"class":219},"\"720p\"",[183,297,298],{"class":193},",       ",[183,300,302],{"class":301},"sAwPA","# 480p o 720p\n",[183,304,306,309,311,315,318],{"class":185,"line":305},10,[183,307,308],{"class":219},"        \"duration\"",[183,310,241],{"class":193},[183,312,314],{"class":313},"sDLfK","5",[183,316,317],{"class":193},",           ",[183,319,320],{"class":301},"# 4–15 segundos (entero)\n",[183,322,324,327,329,332],{"class":185,"line":323},11,[183,325,326],{"class":219},"        \"aspect_ratio\"",[183,328,241],{"class":193},[183,330,331],{"class":219},"\"16:9\"",[183,333,223],{"class":193},[183,335,337,340,342],{"class":185,"line":336},12,[183,338,339],{"class":219},"        \"generate_audio\"",[183,341,241],{"class":193},[183,343,344],{"class":313},"True\n",[183,346,348],{"class":185,"line":347},13,[183,349,350],{"class":193},"    }\n",[183,352,354],{"class":185,"line":353},14,[183,355,356],{"class":193},")\n",[78,358,360],{"id":359},"web-search-opcional","Web Search (opcional)",[15,362,363,366],{},[26,364,365],{},"model_params.web_search"," es exclusivo de la familia text-to-video. Cuando está activado, el modelo decide de forma autónoma si buscar contenido reciente en internet, y solo se factura cuando realmente se realiza una búsqueda:",[174,368,370],{"className":176,"code":369,"language":178,"meta":179,"style":179},"json={\n    \"model\": \"seedance-2.0-text-to-video\",\n    \"prompt\": \"An ad for the latest 2026 electric sports car\",\n    \"model_params\": {\"web_search\": True}\n}\n",[26,371,372,381,392,404,423],{"__ignoreMap":179},[183,373,374,377,379],{"class":185,"line":186},[183,375,376],{"class":193},"json",[183,378,210],{"class":189},[183,380,258],{"class":193},[183,382,383,386,388,390],{"class":185,"line":197},[183,384,385],{"class":219},"    \"model\"",[183,387,241],{"class":193},[183,389,269],{"class":219},[183,391,223],{"class":193},[183,393,394,397,399,402],{"class":185,"line":204},[183,395,396],{"class":219},"    \"prompt\"",[183,398,241],{"class":193},[183,400,401],{"class":219},"\"An ad for the latest 2026 electric sports car\"",[183,403,223],{"class":193},[183,405,406,409,412,415,417,420],{"class":185,"line":216},[183,407,408],{"class":219},"    \"model_params\"",[183,410,411],{"class":193},": {",[183,413,414],{"class":219},"\"web_search\"",[183,416,241],{"class":193},[183,418,419],{"class":313},"True",[183,421,422],{"class":193},"}\n",[183,424,425],{"class":185,"line":226},[183,426,422],{"class":193},[15,428,429,430,434],{},"Consulta la referencia completa de parámetros en la ",[31,431,433],{"href":432},"/docs/text-to-video","documentación de la API Text-to-Video",".",[10,436,438],{"id":437},"casos-de-uso","Casos de uso",[90,440,441,447,453,459],{},[46,442,443,446],{},[21,444,445],{},"E-commerce:"," Genera vídeos de escaparate de productos a partir de descripciones",[46,448,449,452],{},[21,450,451],{},"Marketing:"," Crea creatividades publicitarias y contenido para redes sociales a escala",[46,454,455,458],{},[21,456,457],{},"Entretenimiento:"," Prototipado rápido de escenas y storyboards",[46,460,461,464],{},[21,462,463],{},"Educación:"," Ilustra conceptos abstractos con vídeos personalizados",[10,466,468],{"id":467},"relacionado","Relacionado",[90,470,471,477,483,488],{},[46,472,473],{},[31,474,476],{"href":475},"/models/seedance-2-0","Visión general de Seedance 2.0",[46,478,479],{},[31,480,482],{"href":481},"/models/seedance-2-0-image-to-video","Image-to-Video",[46,484,485],{},[31,486,487],{"href":432},"Referencia completa de la API Text-to-Video",[46,489,490],{},[31,491,492],{"href":33},"Modelos Fast",[494,495,496],"style",{},"html pre.shiki code .snl16, html code.shiki .snl16{--shiki-default:#F97583}html pre.shiki code .s95oV, html code.shiki .s95oV{--shiki-default:#E1E4E8}html pre.shiki code .sU2Wk, html code.shiki .sU2Wk{--shiki-default:#9ECBFF}html pre.shiki code .s9osk, html code.shiki .s9osk{--shiki-default:#FFAB70}html pre.shiki code .sAwPA, html code.shiki .sAwPA{--shiki-default:#6A737D}html pre.shiki code .sDLfK, html code.shiki .sDLfK{--shiki-default:#79B8FF}html .default .shiki span {color: var(--shiki-default);background: var(--shiki-default-bg);font-style: var(--shiki-default-font-style);font-weight: var(--shiki-default-font-weight);text-decoration: var(--shiki-default-text-decoration);}html .shiki span {color: var(--shiki-default);background: var(--shiki-default-bg);font-style: var(--shiki-default-font-style);font-weight: var(--shiki-default-font-weight);text-decoration: var(--shiki-default-text-decoration);}",{"title":179,"searchDepth":197,"depth":197,"links":498},[499,500,501,507,510,511],{"id":12,"depth":197,"text":13},{"id":40,"depth":197,"text":41},{"id":75,"depth":197,"text":76,"children":502},[503,504,505,506],{"id":80,"depth":204,"text":81},{"id":87,"depth":204,"text":88},{"id":146,"depth":204,"text":147},{"id":161,"depth":204,"text":162},{"id":171,"depth":197,"text":172,"children":508},[509],{"id":359,"depth":204,"text":360},{"id":437,"depth":197,"text":438},{"id":467,"depth":197,"text":468},"Genera vídeos cinematográficos a partir de descripciones en lenguaje natural usando seedance-2.0-text-to-video. Admite 480p/720p, duraciones de 4 a 15 segundos y aumento opcional con web_search.","md",{"modelName":515,"tags":516,"metrics":520},"Seedance 2.0 Text-to-Video",[517,518,519],"Text-to-Video","Seedance 2.0","Cinematográfico",[521,524,527,530],{"value":522,"label":523},"720p","Resolución máxima",{"value":525,"label":526},"1000","Palabras máximas en inglés",{"value":528,"label":529},"15s","Duración máxima",{"value":531,"label":532},"30-60s","Tiempo de generación","/es/models/seedance-2-0-text-to-video",{"title":5,"description":512},"es/models/seedance-2-0-text-to-video","tYyFxsAMJX6eQXAJ-pgUmv4lj0gcjM2spuLzhV9pDcM",1776086320038]