|
2 | 2 | msgstr "" |
3 | 3 | "Project-Id-Version: basiliskllm\n" |
4 | 4 | "Report-Msgid-Bugs-To: EMAIL@ADDRESS\n" |
5 | | -"POT-Creation-Date: 2025-10-11 13:06+0000\n" |
6 | | -"PO-Revision-Date: 2025-10-11 13:23\n" |
| 5 | +"POT-Creation-Date: 2026-01-17 13:23+0000\n" |
| 6 | +"PO-Revision-Date: 2026-01-17 13:32\n" |
7 | 7 | "Last-Translator: \n" |
8 | 8 | "Language-Team: French\n" |
9 | 9 | "MIME-Version: 1.0\n" |
@@ -1383,48 +1383,56 @@ msgid "Error checking for updates: %s" |
1383 | 1383 | msgstr "Erreur durant la recherche des mises à jour : %s" |
1384 | 1384 |
|
1385 | 1385 | #: basilisk/provider_engine/anthropic_engine.py:99 |
| 1386 | +msgid "Our fastest model with near-frontier intelligenceBest model for complex agents and coding with highest intelligence" |
| 1387 | +msgstr "" |
| 1388 | + |
1386 | 1389 | #: basilisk/provider_engine/anthropic_engine.py:110 |
1387 | | -msgid "Best model for complex agents and coding with highest intelligence" |
| 1390 | +msgid "Our fastest model with near-frontier intelligence" |
1388 | 1391 | msgstr "" |
1389 | 1392 |
|
1390 | 1393 | #: basilisk/provider_engine/anthropic_engine.py:122 |
1391 | 1394 | #: basilisk/provider_engine/anthropic_engine.py:133 |
| 1395 | +msgid "Best model for complex agents and coding with highest intelligence" |
| 1396 | +msgstr "" |
| 1397 | + |
| 1398 | +#: basilisk/provider_engine/anthropic_engine.py:145 |
| 1399 | +#: basilisk/provider_engine/anthropic_engine.py:156 |
1392 | 1400 | msgid "Exceptional model for specialized complex tasks" |
1393 | 1401 | msgstr "" |
1394 | 1402 |
|
1395 | 1403 | #. Translators: This is a model description |
1396 | | -#: basilisk/provider_engine/anthropic_engine.py:144 |
1397 | | -#: basilisk/provider_engine/anthropic_engine.py:153 |
| 1404 | +#: basilisk/provider_engine/anthropic_engine.py:167 |
| 1405 | +#: basilisk/provider_engine/anthropic_engine.py:176 |
1398 | 1406 | msgid "High-performance model" |
1399 | 1407 | msgstr "" |
1400 | 1408 |
|
1401 | 1409 | #. Translators: This is a model description |
1402 | | -#: basilisk/provider_engine/anthropic_engine.py:163 |
1403 | | -#: basilisk/provider_engine/anthropic_engine.py:172 |
| 1410 | +#: basilisk/provider_engine/anthropic_engine.py:186 |
| 1411 | +#: basilisk/provider_engine/anthropic_engine.py:195 |
1404 | 1412 | msgid "Our most capable model" |
1405 | 1413 | msgstr "" |
1406 | 1414 |
|
1407 | | -#: basilisk/provider_engine/anthropic_engine.py:183 |
1408 | | -#: basilisk/provider_engine/anthropic_engine.py:194 |
| 1415 | +#: basilisk/provider_engine/anthropic_engine.py:206 |
| 1416 | +#: basilisk/provider_engine/anthropic_engine.py:217 |
1409 | 1417 | msgid "High-performance model with early extended thinking" |
1410 | 1418 | msgstr "" |
1411 | 1419 |
|
1412 | 1420 | #. Translators: This is a model description |
1413 | | -#: basilisk/provider_engine/anthropic_engine.py:205 |
| 1421 | +#: basilisk/provider_engine/anthropic_engine.py:228 |
1414 | 1422 | msgid "Our previous intelligent model" |
1415 | 1423 | msgstr "" |
1416 | 1424 |
|
1417 | 1425 | #. Translators: This is a model description |
1418 | | -#: basilisk/provider_engine/anthropic_engine.py:214 |
| 1426 | +#: basilisk/provider_engine/anthropic_engine.py:237 |
1419 | 1427 | msgid "Our fastest model" |
1420 | 1428 | msgstr "" |
1421 | 1429 |
|
1422 | 1430 | #. Translators: This is a model description |
1423 | | -#: basilisk/provider_engine/anthropic_engine.py:223 |
| 1431 | +#: basilisk/provider_engine/anthropic_engine.py:246 |
1424 | 1432 | msgid "Powerful model for complex tasks" |
1425 | 1433 | msgstr "" |
1426 | 1434 |
|
1427 | | -#: basilisk/provider_engine/anthropic_engine.py:233 |
| 1435 | +#: basilisk/provider_engine/anthropic_engine.py:256 |
1428 | 1436 | msgid "Fastest and most compact model for near-instant responsiveness" |
1429 | 1437 | msgstr "Modèle le plus rapide et le plus compact pour une réactivité quasi instantanée" |
1430 | 1438 |
|
@@ -1484,88 +1492,60 @@ msgstr "" |
1484 | 1492 | msgid "GPT-5 model used in ChatGPT" |
1485 | 1493 | msgstr "" |
1486 | 1494 |
|
| 1495 | +#: basilisk/provider_engine/openai_engine.py:156 |
| 1496 | +msgid "Version of GPT-5 that produces smarter and more precise responses" |
| 1497 | +msgstr "" |
| 1498 | + |
1487 | 1499 | #. Translators: This is a model description |
1488 | | -#: basilisk/provider_engine/openai_engine.py:155 |
| 1500 | +#: basilisk/provider_engine/openai_engine.py:167 |
1489 | 1501 | msgid "Flagship GPT model for complex tasks" |
1490 | 1502 | msgstr "" |
1491 | 1503 |
|
1492 | 1504 | #. Translators: This is a model description |
1493 | | -#: basilisk/provider_engine/openai_engine.py:165 |
| 1505 | +#: basilisk/provider_engine/openai_engine.py:177 |
1494 | 1506 | msgid "Balanced for intelligence, speed, and cost" |
1495 | 1507 | msgstr "" |
1496 | 1508 |
|
1497 | 1509 | #. Translators: This is a model description |
1498 | | -#: basilisk/provider_engine/openai_engine.py:175 |
| 1510 | +#: basilisk/provider_engine/openai_engine.py:187 |
1499 | 1511 | msgid "Fastest, most cost-effective GPT-4.1 model" |
1500 | 1512 | msgstr "" |
1501 | 1513 |
|
1502 | 1514 | #. Translators: This is a model description |
1503 | | -#: basilisk/provider_engine/openai_engine.py:185 |
| 1515 | +#: basilisk/provider_engine/openai_engine.py:197 |
1504 | 1516 | msgid "Faster, more affordable reasoning model" |
1505 | 1517 | msgstr "" |
1506 | 1518 |
|
1507 | 1519 | #. Translators: This is a model description |
1508 | | -#: basilisk/provider_engine/openai_engine.py:195 |
| 1520 | +#: basilisk/provider_engine/openai_engine.py:207 |
1509 | 1521 | msgid "Our most powerful reasoning model" |
1510 | 1522 | msgstr "" |
1511 | 1523 |
|
1512 | | -#: basilisk/provider_engine/openai_engine.py:207 |
1513 | | -#: basilisk/provider_engine/openai_engine.py:231 |
| 1524 | +#: basilisk/provider_engine/openai_engine.py:219 |
| 1525 | +#: basilisk/provider_engine/openai_engine.py:243 |
1514 | 1526 | msgid "Points to one of the most recent iterations of gpt-4o-mini model" |
1515 | 1527 | msgstr "Pointe vers l'une des itérations les plus récentes du modèle gpt-4o-mini" |
1516 | 1528 |
|
1517 | | -#: basilisk/provider_engine/openai_engine.py:219 |
| 1529 | +#: basilisk/provider_engine/openai_engine.py:231 |
1518 | 1530 | msgid "Dynamic model continuously updated to the current version of GPT-4o in ChatGPT" |
1519 | 1531 | msgstr "Modèle dynamique mis à jour en continu avec la version actuelle de GPT-4 dans ChatGPT" |
1520 | 1532 |
|
1521 | | -#: basilisk/provider_engine/openai_engine.py:243 |
| 1533 | +#: basilisk/provider_engine/openai_engine.py:255 |
1522 | 1534 | msgid "Our most recent small reasoning model, providing high intelligence at the same cost and latency targets of o1-mini. o3-mini also supports key developer features, like Structured Outputs, function calling, Batch API, and more. Like other models in the o-series, it is designed to excel at science, math, and coding tasks." |
1523 | 1535 | msgstr "" |
1524 | 1536 |
|
1525 | | -#: basilisk/provider_engine/openai_engine.py:256 |
| 1537 | +#: basilisk/provider_engine/openai_engine.py:268 |
1526 | 1538 | msgid "Points to the most recent snapshot of the o1 model" |
1527 | 1539 | msgstr "Pointe vers l'instantané le plus récent du modèle o1" |
1528 | 1540 |
|
1529 | | -#: basilisk/provider_engine/openai_engine.py:269 |
| 1541 | +#: basilisk/provider_engine/openai_engine.py:281 |
1530 | 1542 | msgid "The latest GPT-4 Turbo model with vision capabilities" |
1531 | 1543 | msgstr "Le dernier modèle GPT-4 Turbo avec des capacités de vision" |
1532 | 1544 |
|
1533 | | -#: basilisk/provider_engine/openai_engine.py:281 |
| 1545 | +#: basilisk/provider_engine/openai_engine.py:293 |
1534 | 1546 | msgid "Points to one of the most recent iterations of gpt-3.5 model" |
1535 | 1547 | msgstr "Pointe vers l'une des itérations les plus récentes du modèle gpt-3.5" |
1536 | 1548 |
|
1537 | | -#: basilisk/provider_engine/openai_engine.py:291 |
1538 | | -msgid "Latest snapshot that supports Structured Outputs" |
1539 | | -msgstr "Dernière version prenant en charge les sorties structurées" |
1540 | | - |
1541 | | -#: basilisk/provider_engine/openai_engine.py:303 |
1542 | | -msgid "Our high-intelligence flagship model for complex, multi-step tasks" |
1543 | | -msgstr "Notre modèle phare à haute intelligence pour des tâches complexes et multilingues" |
1544 | | - |
1545 | | -#: basilisk/provider_engine/openai_engine.py:315 |
1546 | | -msgid "Our affordable and intelligent small model for fast, lightweight tasks. GPT-4o mini is cheaper and more capable than GPT-3.5 Turbo" |
1547 | | -msgstr "Notre modèle petit, abordable et intelligent pour des tâches rapides et légères. GPT-4o mini est moins cher et plus performant que GPT-3.5 Turbo" |
1548 | | - |
1549 | | -#: basilisk/provider_engine/openai_engine.py:327 |
1550 | | -msgid "The latest GPT-3.5 Turbo model with higher accuracy at responding in requested formats and a fix for a bug which caused a text encoding issue for non-English language function calls" |
1551 | | -msgstr "Le dernier GPT-3.5 Turbo ProviderAIModel avec une précision accrue dans les réponses aux formats demandés et une correction d’un bogue qui causait un problème d’encodage de texte pour les appels de fonction en langues autres que l’anglais" |
1552 | | - |
1553 | | -#: basilisk/provider_engine/openai_engine.py:337 |
1554 | | -msgid "Points to one of the most recent iterations of gpt-4 model" |
1555 | | -msgstr "Pointe vers l'une des itérations les plus récentes du modèle GPT-4" |
1556 | | - |
1557 | | -#: basilisk/provider_engine/openai_engine.py:348 |
1558 | | -msgid "The latest GPT-4 model intended to reduce cases of “laziness” where the model doesn’t complete a task" |
1559 | | -msgstr "Le dernier modèle GPT-4 ProviderAIModel conçu pour réduire les cas de \"paresse\" où le ProviderAIModel ne termine pas une tâche" |
1560 | | - |
1561 | | -#: basilisk/provider_engine/openai_engine.py:359 |
1562 | | -msgid "GPT-4 Turbo model featuring improved instruction following, JSON mode, reproducible outputs, parallel function calling, and more" |
1563 | | -msgstr "GPT-4 Turbo ProviderAIModel offrant un meilleur suivi des instructions, le mode JSON, des sorties reproductibles, l’appel de fonctions parallèles, et plus encore" |
1564 | | - |
1565 | | -#: basilisk/provider_engine/openai_engine.py:370 |
1566 | | -msgid "More capable than any GPT-3.5 model, able to do more complex tasks, and optimized for chat" |
1567 | | -msgstr "Plus performant que n'importe quel modèle GPT-3.5, capable de réaliser des tâches plus complexes et optimisé pour les conversations" |
1568 | | - |
1569 | 1549 | #: basilisk/provider_engine/xai_engine.py:47 |
1570 | 1550 | msgid "The most intelligent model from xAI, featuring native tool use, real-time search integration, and a context window of 256,000 tokens." |
1571 | 1551 | msgstr "" |
@@ -1601,3 +1581,4 @@ msgstr "" |
1601 | 1581 | #: basilisk/provider_engine/xai_engine.py:130 |
1602 | 1582 | msgid "The original Grok model, providing foundational AI capabilities with a context window of 8,192 tokens." |
1603 | 1583 | msgstr "" |
| 1584 | + |
0 commit comments