[{"data":1,"prerenderedAt":810},["ShallowReactive",2],{"docs-/es/docs/fast-models":3},{"id":4,"title":5,"body":6,"description":803,"extension":804,"meta":805,"navigation":373,"path":806,"seo":807,"stem":808,"__hash__":809},"content/es/docs/fast-models.md","Modelos Fast - Seedance 2.0",{"type":7,"value":8,"toc":793},"minimark",[9,14,35,40,109,112,122,126,131,173,178,195,199,202,208,215,219,339,348,352,648,652,734,738,758,762,789],[10,11,13],"h1",{"id":12},"modelos-fast","Modelos Fast",[15,16,17,18,22,23,26,27,30,31,34],"p",{},"Cada modo de Seedance 2.0 se distribuye en dos niveles: ",[19,20,21],"strong",{},"Estándar"," y ",[19,24,25],{},"Fast",". La familia Fast cede una pequeña cantidad de calidad visual a cambio de una ",[19,28,29],{},"generación más rápida"," y un ",[19,32,33],{},"precio por segundo más bajo",", perfecta para iteración rápida, producción masiva y A/B testing.",[36,37,39],"h2",{"id":38},"los-tres-modelos-fast","Los tres modelos Fast",[41,42,43,59],"table",{},[44,45,46],"thead",{},[47,48,49,53,56],"tr",{},[50,51,52],"th",{},"ID del modelo Fast",[50,54,55],{},"Estándar correspondiente",[50,57,58],{},"Recomendado para",[60,61,62,79,94],"tbody",{},[47,63,64,71,76],{},[65,66,67],"td",{},[68,69,70],"code",{},"seedance-2.0-fast-text-to-video",[65,72,73],{},[68,74,75],{},"seedance-2.0-text-to-video",[65,77,78],{},"Text-to-video puro",[47,80,81,86,91],{},[65,82,83],{},[68,84,85],{},"seedance-2.0-fast-image-to-video",[65,87,88],{},[68,89,90],{},"seedance-2.0-image-to-video",[65,92,93],{},"Conducción con 1 o 2 imágenes",[47,95,96,101,106],{},[65,97,98],{},[68,99,100],{},"seedance-2.0-fast-reference-to-video",[65,102,103],{},[68,104,105],{},"seedance-2.0-reference-to-video",[65,107,108],{},"Composición multimodal",[15,110,111],{},"Todos los modelos Fast comparten el mismo endpoint y la misma estructura de parámetros:",[113,114,119],"pre",{"className":115,"code":117,"language":118},[116],"language-text","POST https://api.evolink.ai/v1/videos/generations\n","text",[68,120,117],{"__ignoreMap":121},"",[36,123,125],{"id":124},"diferencias-respecto-al-estándar","Diferencias respecto al Estándar",[15,127,128],{},[19,129,130],{},"Idéntico:",[132,133,134,138,141,160,163,166],"ul",{},[135,136,137],"li",{},"Endpoint",[135,139,140],{},"Esquema del cuerpo de la solicitud (todos los nombres de parámetros, tipos y valores por defecto)",[135,142,143,144,147,148,151,152,155,156,159],{},"Niveles de calidad permitidos (",[68,145,146],{},"480p"," / ",[68,149,150],{},"720p","), rango de duración (",[68,153,154],{},"4","–",[68,157,158],{},"15"," segundos), relaciones de aspecto",[135,161,162],{},"Cantidad y formato de los activos de entrada",[135,164,165],{},"Esquema de respuesta, ciclo de vida de tareas y formato del payload del webhook",[135,167,168,169,172],{},"Límite de longitud del ",[68,170,171],{},"prompt"," (500 caracteres chinos / 1000 palabras en inglés)",[15,174,175],{},[19,176,177],{},"Diferente:",[132,179,180,183,186,189],{},[135,181,182],{},"Generación más rápida",[135,184,185],{},"Precio por segundo más bajo",[135,187,188],{},"Detalle de calidad ligeramente inferior al Estándar (normalmente indistinguible a simple vista)",[135,190,191,194],{},[68,192,193],{},"fast-image-to-video"," detecta automáticamente el modo (primer fotograma vs primer-último fotograma) según el número de imágenes (1 = conducido por primer fotograma, 2 = transición primer-último fotograma); no se necesita ningún campo adicional",[36,196,198],{"id":197},"flujo-de-trabajo-recomendado","Flujo de trabajo recomendado",[15,200,201],{},"Una pipeline de producción típica utiliza ambos niveles:",[113,203,206],{"className":204,"code":205,"language":118},[116],"Iteración de prompts / parámetros\n    ↓  (Modelo Fast — rápido y barato)\n    ↓\nIdentifica los prompts y parámetros con los que estás satisfecho\n    ↓  (Cambia el campo `model`, deja todo lo demás igual)\n    ↓\nEntrega final\n    ↓  (Modelo Estándar — render de la máxima calidad)\n",[68,207,205],{"__ignoreMap":121},[15,209,210,211,214],{},"A nivel de código, el único cambio es la cadena ",[68,212,213],{},"model",": no hay que tocar ninguna otra lógica.",[36,216,218],{"id":217},"ejemplo-fast-text-to-video-web_search","Ejemplo: Fast Text-to-Video + web_search",[113,220,224],{"className":221,"code":222,"language":223,"meta":121,"style":121},"language-bash shiki shiki-themes github-dark","curl -X POST https://api.evolink.ai/v1/videos/generations \\\n  -H \"Authorization: Bearer YOUR_API_KEY\" \\\n  -H \"Content-Type: application/json\" \\\n  -d '{\n    \"model\": \"seedance-2.0-fast-text-to-video\",\n    \"prompt\": \"A commercial introducing the latest 2026 electric sports car, highlighting its aerodynamic design and cabin tech.\",\n    \"duration\": 6,\n    \"quality\": \"720p\",\n    \"aspect_ratio\": \"16:9\",\n    \"generate_audio\": true,\n    \"model_params\": {\n      \"web_search\": true\n    }\n  }'\n","bash",[68,225,226,249,260,270,279,285,291,297,303,309,315,321,327,333],{"__ignoreMap":121},[227,228,231,235,239,243,246],"span",{"class":229,"line":230},"line",1,[227,232,234],{"class":233},"svObZ","curl",[227,236,238],{"class":237},"sDLfK"," -X",[227,240,242],{"class":241},"sU2Wk"," POST",[227,244,245],{"class":241}," https://api.evolink.ai/v1/videos/generations",[227,247,248],{"class":237}," \\\n",[227,250,252,255,258],{"class":229,"line":251},2,[227,253,254],{"class":237},"  -H",[227,256,257],{"class":241}," \"Authorization: Bearer YOUR_API_KEY\"",[227,259,248],{"class":237},[227,261,263,265,268],{"class":229,"line":262},3,[227,264,254],{"class":237},[227,266,267],{"class":241}," \"Content-Type: application/json\"",[227,269,248],{"class":237},[227,271,273,276],{"class":229,"line":272},4,[227,274,275],{"class":237},"  -d",[227,277,278],{"class":241}," '{\n",[227,280,282],{"class":229,"line":281},5,[227,283,284],{"class":241},"    \"model\": \"seedance-2.0-fast-text-to-video\",\n",[227,286,288],{"class":229,"line":287},6,[227,289,290],{"class":241},"    \"prompt\": \"A commercial introducing the latest 2026 electric sports car, highlighting its aerodynamic design and cabin tech.\",\n",[227,292,294],{"class":229,"line":293},7,[227,295,296],{"class":241},"    \"duration\": 6,\n",[227,298,300],{"class":229,"line":299},8,[227,301,302],{"class":241},"    \"quality\": \"720p\",\n",[227,304,306],{"class":229,"line":305},9,[227,307,308],{"class":241},"    \"aspect_ratio\": \"16:9\",\n",[227,310,312],{"class":229,"line":311},10,[227,313,314],{"class":241},"    \"generate_audio\": true,\n",[227,316,318],{"class":229,"line":317},11,[227,319,320],{"class":241},"    \"model_params\": {\n",[227,322,324],{"class":229,"line":323},12,[227,325,326],{"class":241},"      \"web_search\": true\n",[227,328,330],{"class":229,"line":329},13,[227,331,332],{"class":241},"    }\n",[227,334,336],{"class":229,"line":335},14,[227,337,338],{"class":241},"  }'\n",[340,341,342],"blockquote",{},[15,343,344,347],{},[68,345,346],{},"model_params.web_search"," es exclusivo de la familia text-to-video (incluida la variante Fast). Solo se factura cuando realmente se realiza una búsqueda.",[36,349,351],{"id":350},"ejemplo-fast-image-to-video-auto-primer-fotograma-primer-último-fotograma","Ejemplo: Fast Image-to-Video (auto primer fotograma / primer-último fotograma)",[113,353,357],{"className":354,"code":355,"language":356,"meta":121,"style":121},"language-python shiki shiki-themes github-dark","import requests\n\n# 1 imagen → conducido por primer fotograma\nresponse = requests.post(\n    \"https://api.evolink.ai/v1/videos/generations\",\n    headers={\n        \"Authorization\": \"Bearer YOUR_API_KEY\",\n        \"Content-Type\": \"application/json\"\n    },\n    json={\n        \"model\": \"seedance-2.0-fast-image-to-video\",\n        \"prompt\": \"Camera slowly pushes in, the scene comes alive\",\n        \"image_urls\": [\"https://example.com/scene.jpg\"],\n        \"duration\": 5\n    }\n)\n\n# 2 imágenes → cambia automáticamente a transición primer-último fotograma\nresponse = requests.post(\n    \"https://api.evolink.ai/v1/videos/generations\",\n    headers={\n        \"Authorization\": \"Bearer YOUR_API_KEY\",\n        \"Content-Type\": \"application/json\"\n    },\n    json={\n        \"model\": \"seedance-2.0-fast-image-to-video\",\n        \"prompt\": \"A smooth transition between two scenes\",\n        \"image_urls\": [\n            \"https://example.com/first.jpg\",\n            \"https://example.com/last.jpg\"\n        ],\n        \"duration\": 6\n    }\n)\n","python",[68,358,359,369,375,381,392,400,411,424,434,439,448,460,472,486,496,501,507,512,518,527,534,543,554,563,568,577,588,600,608,616,622,628,638,643],{"__ignoreMap":121},[227,360,361,365],{"class":229,"line":230},[227,362,364],{"class":363},"snl16","import",[227,366,368],{"class":367},"s95oV"," requests\n",[227,370,371],{"class":229,"line":251},[227,372,374],{"emptyLinePlaceholder":373},true,"\n",[227,376,377],{"class":229,"line":262},[227,378,380],{"class":379},"sAwPA","# 1 imagen → conducido por primer fotograma\n",[227,382,383,386,389],{"class":229,"line":272},[227,384,385],{"class":367},"response ",[227,387,388],{"class":363},"=",[227,390,391],{"class":367}," requests.post(\n",[227,393,394,397],{"class":229,"line":281},[227,395,396],{"class":241},"    \"https://api.evolink.ai/v1/videos/generations\"",[227,398,399],{"class":367},",\n",[227,401,402,406,408],{"class":229,"line":287},[227,403,405],{"class":404},"s9osk","    headers",[227,407,388],{"class":363},[227,409,410],{"class":367},"{\n",[227,412,413,416,419,422],{"class":229,"line":293},[227,414,415],{"class":241},"        \"Authorization\"",[227,417,418],{"class":367},": ",[227,420,421],{"class":241},"\"Bearer YOUR_API_KEY\"",[227,423,399],{"class":367},[227,425,426,429,431],{"class":229,"line":299},[227,427,428],{"class":241},"        \"Content-Type\"",[227,430,418],{"class":367},[227,432,433],{"class":241},"\"application/json\"\n",[227,435,436],{"class":229,"line":305},[227,437,438],{"class":367},"    },\n",[227,440,441,444,446],{"class":229,"line":311},[227,442,443],{"class":404},"    json",[227,445,388],{"class":363},[227,447,410],{"class":367},[227,449,450,453,455,458],{"class":229,"line":317},[227,451,452],{"class":241},"        \"model\"",[227,454,418],{"class":367},[227,456,457],{"class":241},"\"seedance-2.0-fast-image-to-video\"",[227,459,399],{"class":367},[227,461,462,465,467,470],{"class":229,"line":323},[227,463,464],{"class":241},"        \"prompt\"",[227,466,418],{"class":367},[227,468,469],{"class":241},"\"Camera slowly pushes in, the scene comes alive\"",[227,471,399],{"class":367},[227,473,474,477,480,483],{"class":229,"line":329},[227,475,476],{"class":241},"        \"image_urls\"",[227,478,479],{"class":367},": [",[227,481,482],{"class":241},"\"https://example.com/scene.jpg\"",[227,484,485],{"class":367},"],\n",[227,487,488,491,493],{"class":229,"line":335},[227,489,490],{"class":241},"        \"duration\"",[227,492,418],{"class":367},[227,494,495],{"class":237},"5\n",[227,497,499],{"class":229,"line":498},15,[227,500,332],{"class":367},[227,502,504],{"class":229,"line":503},16,[227,505,506],{"class":367},")\n",[227,508,510],{"class":229,"line":509},17,[227,511,374],{"emptyLinePlaceholder":373},[227,513,515],{"class":229,"line":514},18,[227,516,517],{"class":379},"# 2 imágenes → cambia automáticamente a transición primer-último fotograma\n",[227,519,521,523,525],{"class":229,"line":520},19,[227,522,385],{"class":367},[227,524,388],{"class":363},[227,526,391],{"class":367},[227,528,530,532],{"class":229,"line":529},20,[227,531,396],{"class":241},[227,533,399],{"class":367},[227,535,537,539,541],{"class":229,"line":536},21,[227,538,405],{"class":404},[227,540,388],{"class":363},[227,542,410],{"class":367},[227,544,546,548,550,552],{"class":229,"line":545},22,[227,547,415],{"class":241},[227,549,418],{"class":367},[227,551,421],{"class":241},[227,553,399],{"class":367},[227,555,557,559,561],{"class":229,"line":556},23,[227,558,428],{"class":241},[227,560,418],{"class":367},[227,562,433],{"class":241},[227,564,566],{"class":229,"line":565},24,[227,567,438],{"class":367},[227,569,571,573,575],{"class":229,"line":570},25,[227,572,443],{"class":404},[227,574,388],{"class":363},[227,576,410],{"class":367},[227,578,580,582,584,586],{"class":229,"line":579},26,[227,581,452],{"class":241},[227,583,418],{"class":367},[227,585,457],{"class":241},[227,587,399],{"class":367},[227,589,591,593,595,598],{"class":229,"line":590},27,[227,592,464],{"class":241},[227,594,418],{"class":367},[227,596,597],{"class":241},"\"A smooth transition between two scenes\"",[227,599,399],{"class":367},[227,601,603,605],{"class":229,"line":602},28,[227,604,476],{"class":241},[227,606,607],{"class":367},": [\n",[227,609,611,614],{"class":229,"line":610},29,[227,612,613],{"class":241},"            \"https://example.com/first.jpg\"",[227,615,399],{"class":367},[227,617,619],{"class":229,"line":618},30,[227,620,621],{"class":241},"            \"https://example.com/last.jpg\"\n",[227,623,625],{"class":229,"line":624},31,[227,626,627],{"class":367},"        ],\n",[227,629,631,633,635],{"class":229,"line":630},32,[227,632,490],{"class":241},[227,634,418],{"class":367},[227,636,637],{"class":237},"6\n",[227,639,641],{"class":229,"line":640},33,[227,642,332],{"class":367},[227,644,646],{"class":229,"line":645},34,[227,647,506],{"class":367},[36,649,651],{"id":650},"ejemplo-fast-reference-to-video","Ejemplo: Fast Reference-to-Video",[113,653,655],{"className":221,"code":654,"language":223,"meta":121,"style":121},"curl -X POST https://api.evolink.ai/v1/videos/generations \\\n  -H \"Authorization: Bearer YOUR_API_KEY\" \\\n  -H \"Content-Type: application/json\" \\\n  -d '{\n    \"model\": \"seedance-2.0-fast-reference-to-video\",\n    \"prompt\": \"Replicate video 1 first-person perspective. Use audio 1 as background music throughout. Promo video opening.\",\n    \"image_urls\": [\"https://example.com/ref1.jpg\"],\n    \"video_urls\": [\"https://example.com/reference.mp4\"],\n    \"audio_urls\": [\"https://example.com/bgm.mp3\"],\n    \"duration\": 10,\n    \"quality\": \"720p\",\n    \"aspect_ratio\": \"16:9\"\n  }'\n",[68,656,657,669,677,685,691,696,701,706,711,716,721,725,730],{"__ignoreMap":121},[227,658,659,661,663,665,667],{"class":229,"line":230},[227,660,234],{"class":233},[227,662,238],{"class":237},[227,664,242],{"class":241},[227,666,245],{"class":241},[227,668,248],{"class":237},[227,670,671,673,675],{"class":229,"line":251},[227,672,254],{"class":237},[227,674,257],{"class":241},[227,676,248],{"class":237},[227,678,679,681,683],{"class":229,"line":262},[227,680,254],{"class":237},[227,682,267],{"class":241},[227,684,248],{"class":237},[227,686,687,689],{"class":229,"line":272},[227,688,275],{"class":237},[227,690,278],{"class":241},[227,692,693],{"class":229,"line":281},[227,694,695],{"class":241},"    \"model\": \"seedance-2.0-fast-reference-to-video\",\n",[227,697,698],{"class":229,"line":287},[227,699,700],{"class":241},"    \"prompt\": \"Replicate video 1 first-person perspective. Use audio 1 as background music throughout. Promo video opening.\",\n",[227,702,703],{"class":229,"line":293},[227,704,705],{"class":241},"    \"image_urls\": [\"https://example.com/ref1.jpg\"],\n",[227,707,708],{"class":229,"line":299},[227,709,710],{"class":241},"    \"video_urls\": [\"https://example.com/reference.mp4\"],\n",[227,712,713],{"class":229,"line":305},[227,714,715],{"class":241},"    \"audio_urls\": [\"https://example.com/bgm.mp3\"],\n",[227,717,718],{"class":229,"line":311},[227,719,720],{"class":241},"    \"duration\": 10,\n",[227,722,723],{"class":229,"line":317},[227,724,302],{"class":241},[227,726,727],{"class":229,"line":323},[227,728,729],{"class":241},"    \"aspect_ratio\": \"16:9\"\n",[227,731,732],{"class":229,"line":329},[227,733,338],{"class":241},[36,735,737],{"id":736},"cuándo-no-usar-fast","Cuándo no usar Fast",[132,739,740,746,752],{},[135,741,742,745],{},[19,743,744],{},"Entregables publicitarios finales / vídeos hero de marca",": elige Estándar para mayor estabilidad de detalle",[135,747,748,751],{},[19,749,750],{},"Primeros planos faciales o microexpresiones",": Estándar es más preciso",[135,753,754,757],{},[19,755,756],{},"Composiciones reference-to-video complejas"," con 9 imágenes + 3 vídeos + 3 audios: Estándar comprende mejor las señales combinadas",[36,759,761],{"id":760},"relacionado","Relacionado",[132,763,764,771,777,783],{},[135,765,766],{},[767,768,770],"a",{"href":769},"/docs/models-overview","Visión general de modelos",[135,772,773],{},[767,774,776],{"href":775},"/docs/text-to-video","API Text-to-Video",[135,778,779],{},[767,780,782],{"href":781},"/docs/image-to-video","API Image-to-Video",[135,784,785],{},[767,786,788],{"href":787},"/docs/reference-to-video","API Reference-to-Video",[790,791,792],"style",{},"html pre.shiki code .svObZ, html code.shiki .svObZ{--shiki-default:#B392F0}html pre.shiki code .sDLfK, html code.shiki .sDLfK{--shiki-default:#79B8FF}html pre.shiki code .sU2Wk, html code.shiki .sU2Wk{--shiki-default:#9ECBFF}html .default .shiki span {color: var(--shiki-default);background: var(--shiki-default-bg);font-style: var(--shiki-default-font-style);font-weight: var(--shiki-default-font-weight);text-decoration: var(--shiki-default-text-decoration);}html .shiki span {color: var(--shiki-default);background: var(--shiki-default-bg);font-style: var(--shiki-default-font-style);font-weight: var(--shiki-default-font-weight);text-decoration: var(--shiki-default-text-decoration);}html pre.shiki code .snl16, html code.shiki .snl16{--shiki-default:#F97583}html pre.shiki code .s95oV, html code.shiki .s95oV{--shiki-default:#E1E4E8}html pre.shiki code .sAwPA, html code.shiki .sAwPA{--shiki-default:#6A737D}html pre.shiki code .s9osk, html code.shiki .s9osk{--shiki-default:#FFAB70}",{"title":121,"searchDepth":251,"depth":251,"links":794},[795,796,797,798,799,800,801,802],{"id":38,"depth":251,"text":39},{"id":124,"depth":251,"text":125},{"id":197,"depth":251,"text":198},{"id":217,"depth":251,"text":218},{"id":350,"depth":251,"text":351},{"id":650,"depth":251,"text":651},{"id":736,"depth":251,"text":737},{"id":760,"depth":251,"text":761},"Familia Fast de Seedance 2.0 (fast-text-to-video / fast-image-to-video / fast-reference-to-video): generación más rápida, menor coste por segundo, misma forma de API.","md",{},"/es/docs/fast-models",{"title":5,"description":803},"es/docs/fast-models","igQfj_I2b_j2jsin1zXk0WmjlXqrzLzVssfpG9M9AyU",1776086323165]