2025-08-11 11:26:15 -07:00
|
|
|
|
use crate::colors::LIGHT_BLUE;
|
2025-08-11 12:31:34 -07:00
|
|
|
|
use crate::diff_render::create_diff_summary;
|
2025-08-06 14:36:48 -07:00
|
|
|
|
use crate::exec_command::relativize_to_home;
|
2025-07-31 00:43:21 -07:00
|
|
|
|
use crate::exec_command::strip_bash_lc_and_escape;
|
2025-08-06 14:36:48 -07:00
|
|
|
|
use crate::slash_command::SlashCommand;
|
2025-06-03 14:29:26 -07:00
|
|
|
|
use crate::text_formatting::format_and_truncate_tool_result;
|
fix: introduce ResponseInputItem::McpToolCallOutput variant (#1151)
The output of an MCP server tool call can be one of several types, but
to date, we treated all outputs as text by showing the serialized JSON
as the "tool output" in Codex:
https://github.com/openai/codex/blob/25a9949c49194d5a64de54a11bcc5b4724ac9bd5/codex-rs/mcp-types/src/lib.rs#L96-L101
This PR adds support for the `ImageContent` variant so we can now
display an image output from an MCP tool call.
In making this change, we introduce a new
`ResponseInputItem::McpToolCallOutput` variant so that we can work with
the `mcp_types::CallToolResult` directly when the function call is made
to an MCP server.
Though arguably the more significant change is the introduction of
`HistoryCell::CompletedMcpToolCallWithImageOutput`, which is a cell that
uses `ratatui_image` to render an image into the terminal. To support
this, we introduce `ImageRenderCache`, cache a
`ratatui_image::picker::Picker`, and `ensure_image_cache()` to cache the
appropriate scaled image data and dimensions based on the current
terminal size.
To test, I created a minimal `package.json`:
```json
{
"name": "kitty-mcp",
"version": "1.0.0",
"type": "module",
"description": "MCP that returns image of kitty",
"main": "index.js",
"dependencies": {
"@modelcontextprotocol/sdk": "^1.12.0"
}
}
```
with the following `index.js` to define the MCP server:
```js
#!/usr/bin/env node
import { McpServer } from "@modelcontextprotocol/sdk/server/mcp.js";
import { StdioServerTransport } from "@modelcontextprotocol/sdk/server/stdio.js";
import { readFile } from "node:fs/promises";
import { join } from "node:path";
const IMAGE_URI = "image://Ada.png";
const server = new McpServer({
name: "Demo",
version: "1.0.0",
});
server.tool(
"get-cat-image",
"If you need a cat image, this tool will provide one.",
async () => ({
content: [
{ type: "image", data: await getAdaPngBase64(), mimeType: "image/png" },
],
})
);
server.resource("Ada the Cat", IMAGE_URI, async (uri) => {
const base64Image = await getAdaPngBase64();
return {
contents: [
{
uri: uri.href,
mimeType: "image/png",
blob: base64Image,
},
],
};
});
async function getAdaPngBase64() {
const __dirname = new URL(".", import.meta.url).pathname;
// From https://github.com/benjajaja/ratatui-image/blob/9705ce2c59ec669abbce2924cbfd1f5ae22c9860/assets/Ada.png
const filePath = join(__dirname, "Ada.png");
const imageData = await readFile(filePath);
const base64Image = imageData.toString("base64");
return base64Image;
}
const transport = new StdioServerTransport();
await server.connect(transport);
```
With the local changes from this PR, I added the following to my
`config.toml`:
```toml
[mcp_servers.kitty]
command = "node"
args = ["/Users/mbolin/code/kitty-mcp/index.js"]
```
Running the TUI from source:
```
cargo run --bin codex -- --model o3 'I need a picture of a cat'
```
I get:
<img width="732" alt="image"
src="https://github.com/user-attachments/assets/bf80b721-9ca0-4d81-aec7-77d6899e2869"
/>
Now, that said, I have only tested in iTerm and there is definitely some
funny business with getting an accurate character-to-pixel ratio
(sometimes the `CompletedMcpToolCallWithImageOutput` thinks it needs 10
rows to render instead of 4), so there is still work to be done here.
2025-05-28 19:03:17 -07:00
|
|
|
|
use base64::Engine;
|
feat: initial import of Rust implementation of Codex CLI in codex-rs/ (#629)
As stated in `codex-rs/README.md`:
Today, Codex CLI is written in TypeScript and requires Node.js 22+ to
run it. For a number of users, this runtime requirement inhibits
adoption: they would be better served by a standalone executable. As
maintainers, we want Codex to run efficiently in a wide range of
environments with minimal overhead. We also want to take advantage of
operating system-specific APIs to provide better sandboxing, where
possible.
To that end, we are moving forward with a Rust implementation of Codex
CLI contained in this folder, which has the following benefits:
- The CLI compiles to small, standalone, platform-specific binaries.
- Can make direct, native calls to
[seccomp](https://man7.org/linux/man-pages/man2/seccomp.2.html) and
[landlock](https://man7.org/linux/man-pages/man7/landlock.7.html) in
order to support sandboxing on Linux.
- No runtime garbage collection, resulting in lower memory consumption
and better, more predictable performance.
Currently, the Rust implementation is materially behind the TypeScript
implementation in functionality, so continue to use the TypeScript
implmentation for the time being. We will publish native executables via
GitHub Releases as soon as we feel the Rust version is usable.
2025-04-24 13:31:40 -07:00
|
|
|
|
use codex_ansi_escape::ansi_escape_line;
|
2025-08-05 23:57:52 -07:00
|
|
|
|
use codex_common::create_config_summary_entries;
|
2025-05-06 17:38:56 -07:00
|
|
|
|
use codex_common::elapsed::format_duration;
|
2025-04-27 21:47:50 -07:00
|
|
|
|
use codex_core::config::Config;
|
2025-08-11 11:26:15 -07:00
|
|
|
|
use codex_core::parse_command::ParsedCommand;
|
2025-07-31 13:45:52 -07:00
|
|
|
|
use codex_core::plan_tool::PlanItemArg;
|
|
|
|
|
|
use codex_core::plan_tool::StepStatus;
|
|
|
|
|
|
use codex_core::plan_tool::UpdatePlanArgs;
|
feat: initial import of Rust implementation of Codex CLI in codex-rs/ (#629)
As stated in `codex-rs/README.md`:
Today, Codex CLI is written in TypeScript and requires Node.js 22+ to
run it. For a number of users, this runtime requirement inhibits
adoption: they would be better served by a standalone executable. As
maintainers, we want Codex to run efficiently in a wide range of
environments with minimal overhead. We also want to take advantage of
operating system-specific APIs to provide better sandboxing, where
possible.
To that end, we are moving forward with a Rust implementation of Codex
CLI contained in this folder, which has the following benefits:
- The CLI compiles to small, standalone, platform-specific binaries.
- Can make direct, native calls to
[seccomp](https://man7.org/linux/man-pages/man2/seccomp.2.html) and
[landlock](https://man7.org/linux/man-pages/man7/landlock.7.html) in
order to support sandboxing on Linux.
- No runtime garbage collection, resulting in lower memory consumption
and better, more predictable performance.
Currently, the Rust implementation is materially behind the TypeScript
implementation in functionality, so continue to use the TypeScript
implmentation for the time being. We will publish native executables via
GitHub Releases as soon as we feel the Rust version is usable.
2025-04-24 13:31:40 -07:00
|
|
|
|
use codex_core::protocol::FileChange;
|
2025-07-30 10:05:40 -07:00
|
|
|
|
use codex_core::protocol::McpInvocation;
|
2025-08-07 04:02:58 -07:00
|
|
|
|
use codex_core::protocol::SandboxPolicy;
|
2025-05-13 19:22:16 -07:00
|
|
|
|
use codex_core::protocol::SessionConfiguredEvent;
|
2025-08-05 23:57:52 -07:00
|
|
|
|
use codex_core::protocol::TokenUsage;
|
2025-08-07 01:27:45 -07:00
|
|
|
|
use codex_login::get_auth_file;
|
|
|
|
|
|
use codex_login::try_read_auth_json;
|
fix: introduce ResponseInputItem::McpToolCallOutput variant (#1151)
The output of an MCP server tool call can be one of several types, but
to date, we treated all outputs as text by showing the serialized JSON
as the "tool output" in Codex:
https://github.com/openai/codex/blob/25a9949c49194d5a64de54a11bcc5b4724ac9bd5/codex-rs/mcp-types/src/lib.rs#L96-L101
This PR adds support for the `ImageContent` variant so we can now
display an image output from an MCP tool call.
In making this change, we introduce a new
`ResponseInputItem::McpToolCallOutput` variant so that we can work with
the `mcp_types::CallToolResult` directly when the function call is made
to an MCP server.
Though arguably the more significant change is the introduction of
`HistoryCell::CompletedMcpToolCallWithImageOutput`, which is a cell that
uses `ratatui_image` to render an image into the terminal. To support
this, we introduce `ImageRenderCache`, cache a
`ratatui_image::picker::Picker`, and `ensure_image_cache()` to cache the
appropriate scaled image data and dimensions based on the current
terminal size.
To test, I created a minimal `package.json`:
```json
{
"name": "kitty-mcp",
"version": "1.0.0",
"type": "module",
"description": "MCP that returns image of kitty",
"main": "index.js",
"dependencies": {
"@modelcontextprotocol/sdk": "^1.12.0"
}
}
```
with the following `index.js` to define the MCP server:
```js
#!/usr/bin/env node
import { McpServer } from "@modelcontextprotocol/sdk/server/mcp.js";
import { StdioServerTransport } from "@modelcontextprotocol/sdk/server/stdio.js";
import { readFile } from "node:fs/promises";
import { join } from "node:path";
const IMAGE_URI = "image://Ada.png";
const server = new McpServer({
name: "Demo",
version: "1.0.0",
});
server.tool(
"get-cat-image",
"If you need a cat image, this tool will provide one.",
async () => ({
content: [
{ type: "image", data: await getAdaPngBase64(), mimeType: "image/png" },
],
})
);
server.resource("Ada the Cat", IMAGE_URI, async (uri) => {
const base64Image = await getAdaPngBase64();
return {
contents: [
{
uri: uri.href,
mimeType: "image/png",
blob: base64Image,
},
],
};
});
async function getAdaPngBase64() {
const __dirname = new URL(".", import.meta.url).pathname;
// From https://github.com/benjajaja/ratatui-image/blob/9705ce2c59ec669abbce2924cbfd1f5ae22c9860/assets/Ada.png
const filePath = join(__dirname, "Ada.png");
const imageData = await readFile(filePath);
const base64Image = imageData.toString("base64");
return base64Image;
}
const transport = new StdioServerTransport();
await server.connect(transport);
```
With the local changes from this PR, I added the following to my
`config.toml`:
```toml
[mcp_servers.kitty]
command = "node"
args = ["/Users/mbolin/code/kitty-mcp/index.js"]
```
Running the TUI from source:
```
cargo run --bin codex -- --model o3 'I need a picture of a cat'
```
I get:
<img width="732" alt="image"
src="https://github.com/user-attachments/assets/bf80b721-9ca0-4d81-aec7-77d6899e2869"
/>
Now, that said, I have only tested in iTerm and there is definitely some
funny business with getting an accurate character-to-pixel ratio
(sometimes the `CompletedMcpToolCallWithImageOutput` thinks it needs 10
rows to render instead of 4), so there is still work to be done here.
2025-05-28 19:03:17 -07:00
|
|
|
|
use image::DynamicImage;
|
|
|
|
|
|
use image::ImageReader;
|
2025-06-03 14:29:26 -07:00
|
|
|
|
use mcp_types::EmbeddedResourceResource;
|
2025-07-19 00:09:34 -04:00
|
|
|
|
use mcp_types::ResourceLink;
|
feat: initial import of Rust implementation of Codex CLI in codex-rs/ (#629)
As stated in `codex-rs/README.md`:
Today, Codex CLI is written in TypeScript and requires Node.js 22+ to
run it. For a number of users, this runtime requirement inhibits
adoption: they would be better served by a standalone executable. As
maintainers, we want Codex to run efficiently in a wide range of
environments with minimal overhead. We also want to take advantage of
operating system-specific APIs to provide better sandboxing, where
possible.
To that end, we are moving forward with a Rust implementation of Codex
CLI contained in this folder, which has the following benefits:
- The CLI compiles to small, standalone, platform-specific binaries.
- Can make direct, native calls to
[seccomp](https://man7.org/linux/man-pages/man2/seccomp.2.html) and
[landlock](https://man7.org/linux/man-pages/man7/landlock.7.html) in
order to support sandboxing on Linux.
- No runtime garbage collection, resulting in lower memory consumption
and better, more predictable performance.
Currently, the Rust implementation is materially behind the TypeScript
implementation in functionality, so continue to use the TypeScript
implmentation for the time being. We will publish native executables via
GitHub Releases as soon as we feel the Rust version is usable.
2025-04-24 13:31:40 -07:00
|
|
|
|
use ratatui::prelude::*;
|
|
|
|
|
|
use ratatui::style::Color;
|
|
|
|
|
|
use ratatui::style::Modifier;
|
|
|
|
|
|
use ratatui::style::Style;
|
2025-08-06 12:03:45 -07:00
|
|
|
|
use ratatui::widgets::Paragraph;
|
|
|
|
|
|
use ratatui::widgets::WidgetRef;
|
2025-08-07 18:38:39 -07:00
|
|
|
|
use ratatui::widgets::Wrap;
|
feat: initial import of Rust implementation of Codex CLI in codex-rs/ (#629)
As stated in `codex-rs/README.md`:
Today, Codex CLI is written in TypeScript and requires Node.js 22+ to
run it. For a number of users, this runtime requirement inhibits
adoption: they would be better served by a standalone executable. As
maintainers, we want Codex to run efficiently in a wide range of
environments with minimal overhead. We also want to take advantage of
operating system-specific APIs to provide better sandboxing, where
possible.
To that end, we are moving forward with a Rust implementation of Codex
CLI contained in this folder, which has the following benefits:
- The CLI compiles to small, standalone, platform-specific binaries.
- Can make direct, native calls to
[seccomp](https://man7.org/linux/man-pages/man2/seccomp.2.html) and
[landlock](https://man7.org/linux/man-pages/man7/landlock.7.html) in
order to support sandboxing on Linux.
- No runtime garbage collection, resulting in lower memory consumption
and better, more predictable performance.
Currently, the Rust implementation is materially behind the TypeScript
implementation in functionality, so continue to use the TypeScript
implmentation for the time being. We will publish native executables via
GitHub Releases as soon as we feel the Rust version is usable.
2025-04-24 13:31:40 -07:00
|
|
|
|
use std::collections::HashMap;
|
fix: introduce ResponseInputItem::McpToolCallOutput variant (#1151)
The output of an MCP server tool call can be one of several types, but
to date, we treated all outputs as text by showing the serialized JSON
as the "tool output" in Codex:
https://github.com/openai/codex/blob/25a9949c49194d5a64de54a11bcc5b4724ac9bd5/codex-rs/mcp-types/src/lib.rs#L96-L101
This PR adds support for the `ImageContent` variant so we can now
display an image output from an MCP tool call.
In making this change, we introduce a new
`ResponseInputItem::McpToolCallOutput` variant so that we can work with
the `mcp_types::CallToolResult` directly when the function call is made
to an MCP server.
Though arguably the more significant change is the introduction of
`HistoryCell::CompletedMcpToolCallWithImageOutput`, which is a cell that
uses `ratatui_image` to render an image into the terminal. To support
this, we introduce `ImageRenderCache`, cache a
`ratatui_image::picker::Picker`, and `ensure_image_cache()` to cache the
appropriate scaled image data and dimensions based on the current
terminal size.
To test, I created a minimal `package.json`:
```json
{
"name": "kitty-mcp",
"version": "1.0.0",
"type": "module",
"description": "MCP that returns image of kitty",
"main": "index.js",
"dependencies": {
"@modelcontextprotocol/sdk": "^1.12.0"
}
}
```
with the following `index.js` to define the MCP server:
```js
#!/usr/bin/env node
import { McpServer } from "@modelcontextprotocol/sdk/server/mcp.js";
import { StdioServerTransport } from "@modelcontextprotocol/sdk/server/stdio.js";
import { readFile } from "node:fs/promises";
import { join } from "node:path";
const IMAGE_URI = "image://Ada.png";
const server = new McpServer({
name: "Demo",
version: "1.0.0",
});
server.tool(
"get-cat-image",
"If you need a cat image, this tool will provide one.",
async () => ({
content: [
{ type: "image", data: await getAdaPngBase64(), mimeType: "image/png" },
],
})
);
server.resource("Ada the Cat", IMAGE_URI, async (uri) => {
const base64Image = await getAdaPngBase64();
return {
contents: [
{
uri: uri.href,
mimeType: "image/png",
blob: base64Image,
},
],
};
});
async function getAdaPngBase64() {
const __dirname = new URL(".", import.meta.url).pathname;
// From https://github.com/benjajaja/ratatui-image/blob/9705ce2c59ec669abbce2924cbfd1f5ae22c9860/assets/Ada.png
const filePath = join(__dirname, "Ada.png");
const imageData = await readFile(filePath);
const base64Image = imageData.toString("base64");
return base64Image;
}
const transport = new StdioServerTransport();
await server.connect(transport);
```
With the local changes from this PR, I added the following to my
`config.toml`:
```toml
[mcp_servers.kitty]
command = "node"
args = ["/Users/mbolin/code/kitty-mcp/index.js"]
```
Running the TUI from source:
```
cargo run --bin codex -- --model o3 'I need a picture of a cat'
```
I get:
<img width="732" alt="image"
src="https://github.com/user-attachments/assets/bf80b721-9ca0-4d81-aec7-77d6899e2869"
/>
Now, that said, I have only tested in iTerm and there is definitely some
funny business with getting an accurate character-to-pixel ratio
(sometimes the `CompletedMcpToolCallWithImageOutput` thinks it needs 10
rows to render instead of 4), so there is still work to be done here.
2025-05-28 19:03:17 -07:00
|
|
|
|
use std::io::Cursor;
|
feat: initial import of Rust implementation of Codex CLI in codex-rs/ (#629)
As stated in `codex-rs/README.md`:
Today, Codex CLI is written in TypeScript and requires Node.js 22+ to
run it. For a number of users, this runtime requirement inhibits
adoption: they would be better served by a standalone executable. As
maintainers, we want Codex to run efficiently in a wide range of
environments with minimal overhead. We also want to take advantage of
operating system-specific APIs to provide better sandboxing, where
possible.
To that end, we are moving forward with a Rust implementation of Codex
CLI contained in this folder, which has the following benefits:
- The CLI compiles to small, standalone, platform-specific binaries.
- Can make direct, native calls to
[seccomp](https://man7.org/linux/man-pages/man2/seccomp.2.html) and
[landlock](https://man7.org/linux/man-pages/man7/landlock.7.html) in
order to support sandboxing on Linux.
- No runtime garbage collection, resulting in lower memory consumption
and better, more predictable performance.
Currently, the Rust implementation is materially behind the TypeScript
implementation in functionality, so continue to use the TypeScript
implmentation for the time being. We will publish native executables via
GitHub Releases as soon as we feel the Rust version is usable.
2025-04-24 13:31:40 -07:00
|
|
|
|
use std::path::PathBuf;
|
|
|
|
|
|
use std::time::Duration;
|
fix: introduce ResponseInputItem::McpToolCallOutput variant (#1151)
The output of an MCP server tool call can be one of several types, but
to date, we treated all outputs as text by showing the serialized JSON
as the "tool output" in Codex:
https://github.com/openai/codex/blob/25a9949c49194d5a64de54a11bcc5b4724ac9bd5/codex-rs/mcp-types/src/lib.rs#L96-L101
This PR adds support for the `ImageContent` variant so we can now
display an image output from an MCP tool call.
In making this change, we introduce a new
`ResponseInputItem::McpToolCallOutput` variant so that we can work with
the `mcp_types::CallToolResult` directly when the function call is made
to an MCP server.
Though arguably the more significant change is the introduction of
`HistoryCell::CompletedMcpToolCallWithImageOutput`, which is a cell that
uses `ratatui_image` to render an image into the terminal. To support
this, we introduce `ImageRenderCache`, cache a
`ratatui_image::picker::Picker`, and `ensure_image_cache()` to cache the
appropriate scaled image data and dimensions based on the current
terminal size.
To test, I created a minimal `package.json`:
```json
{
"name": "kitty-mcp",
"version": "1.0.0",
"type": "module",
"description": "MCP that returns image of kitty",
"main": "index.js",
"dependencies": {
"@modelcontextprotocol/sdk": "^1.12.0"
}
}
```
with the following `index.js` to define the MCP server:
```js
#!/usr/bin/env node
import { McpServer } from "@modelcontextprotocol/sdk/server/mcp.js";
import { StdioServerTransport } from "@modelcontextprotocol/sdk/server/stdio.js";
import { readFile } from "node:fs/promises";
import { join } from "node:path";
const IMAGE_URI = "image://Ada.png";
const server = new McpServer({
name: "Demo",
version: "1.0.0",
});
server.tool(
"get-cat-image",
"If you need a cat image, this tool will provide one.",
async () => ({
content: [
{ type: "image", data: await getAdaPngBase64(), mimeType: "image/png" },
],
})
);
server.resource("Ada the Cat", IMAGE_URI, async (uri) => {
const base64Image = await getAdaPngBase64();
return {
contents: [
{
uri: uri.href,
mimeType: "image/png",
blob: base64Image,
},
],
};
});
async function getAdaPngBase64() {
const __dirname = new URL(".", import.meta.url).pathname;
// From https://github.com/benjajaja/ratatui-image/blob/9705ce2c59ec669abbce2924cbfd1f5ae22c9860/assets/Ada.png
const filePath = join(__dirname, "Ada.png");
const imageData = await readFile(filePath);
const base64Image = imageData.toString("base64");
return base64Image;
}
const transport = new StdioServerTransport();
await server.connect(transport);
```
With the local changes from this PR, I added the following to my
`config.toml`:
```toml
[mcp_servers.kitty]
command = "node"
args = ["/Users/mbolin/code/kitty-mcp/index.js"]
```
Running the TUI from source:
```
cargo run --bin codex -- --model o3 'I need a picture of a cat'
```
I get:
<img width="732" alt="image"
src="https://github.com/user-attachments/assets/bf80b721-9ca0-4d81-aec7-77d6899e2869"
/>
Now, that said, I have only tested in iTerm and there is definitely some
funny business with getting an accurate character-to-pixel ratio
(sometimes the `CompletedMcpToolCallWithImageOutput` thinks it needs 10
rows to render instead of 4), so there is still work to be done here.
2025-05-28 19:03:17 -07:00
|
|
|
|
use tracing::error;
|
feat: initial import of Rust implementation of Codex CLI in codex-rs/ (#629)
As stated in `codex-rs/README.md`:
Today, Codex CLI is written in TypeScript and requires Node.js 22+ to
run it. For a number of users, this runtime requirement inhibits
adoption: they would be better served by a standalone executable. As
maintainers, we want Codex to run efficiently in a wide range of
environments with minimal overhead. We also want to take advantage of
operating system-specific APIs to provide better sandboxing, where
possible.
To that end, we are moving forward with a Rust implementation of Codex
CLI contained in this folder, which has the following benefits:
- The CLI compiles to small, standalone, platform-specific binaries.
- Can make direct, native calls to
[seccomp](https://man7.org/linux/man-pages/man2/seccomp.2.html) and
[landlock](https://man7.org/linux/man-pages/man7/landlock.7.html) in
order to support sandboxing on Linux.
- No runtime garbage collection, resulting in lower memory consumption
and better, more predictable performance.
Currently, the Rust implementation is materially behind the TypeScript
implementation in functionality, so continue to use the TypeScript
implmentation for the time being. We will publish native executables via
GitHub Releases as soon as we feel the Rust version is usable.
2025-04-24 13:31:40 -07:00
|
|
|
|
|
2025-08-11 12:40:12 -07:00
|
|
|
|
#[derive(Clone)]
|
feat: initial import of Rust implementation of Codex CLI in codex-rs/ (#629)
As stated in `codex-rs/README.md`:
Today, Codex CLI is written in TypeScript and requires Node.js 22+ to
run it. For a number of users, this runtime requirement inhibits
adoption: they would be better served by a standalone executable. As
maintainers, we want Codex to run efficiently in a wide range of
environments with minimal overhead. We also want to take advantage of
operating system-specific APIs to provide better sandboxing, where
possible.
To that end, we are moving forward with a Rust implementation of Codex
CLI contained in this folder, which has the following benefits:
- The CLI compiles to small, standalone, platform-specific binaries.
- Can make direct, native calls to
[seccomp](https://man7.org/linux/man-pages/man2/seccomp.2.html) and
[landlock](https://man7.org/linux/man-pages/man7/landlock.7.html) in
order to support sandboxing on Linux.
- No runtime garbage collection, resulting in lower memory consumption
and better, more predictable performance.
Currently, the Rust implementation is materially behind the TypeScript
implementation in functionality, so continue to use the TypeScript
implmentation for the time being. We will publish native executables via
GitHub Releases as soon as we feel the Rust version is usable.
2025-04-24 13:31:40 -07:00
|
|
|
|
pub(crate) struct CommandOutput {
|
|
|
|
|
|
pub(crate) exit_code: i32,
|
|
|
|
|
|
pub(crate) stdout: String,
|
|
|
|
|
|
pub(crate) stderr: String,
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
pub(crate) enum PatchEventType {
|
|
|
|
|
|
ApprovalRequest,
|
|
|
|
|
|
ApplyBegin { auto_approved: bool },
|
|
|
|
|
|
}
|
|
|
|
|
|
|
2025-08-14 14:10:05 -04:00
|
|
|
|
/// Represents an event to display in the conversation history. Returns its
|
|
|
|
|
|
/// `Vec<Line<'static>>` representation to make it easier to display in a
|
|
|
|
|
|
/// scrollable list.
|
|
|
|
|
|
pub(crate) trait HistoryCell {
|
|
|
|
|
|
fn display_lines(&self) -> Vec<Line<'static>>;
|
|
|
|
|
|
|
|
|
|
|
|
fn desired_height(&self, width: u16) -> u16 {
|
|
|
|
|
|
Paragraph::new(Text::from(self.display_lines()))
|
|
|
|
|
|
.wrap(Wrap { trim: false })
|
|
|
|
|
|
.line_count(width)
|
|
|
|
|
|
.try_into()
|
|
|
|
|
|
.unwrap_or(0)
|
2025-07-30 10:05:40 -07:00
|
|
|
|
}
|
|
|
|
|
|
}
|
|
|
|
|
|
|
2025-08-14 14:10:05 -04:00
|
|
|
|
pub(crate) struct PlainHistoryCell {
|
|
|
|
|
|
lines: Vec<Line<'static>>,
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
impl HistoryCell for PlainHistoryCell {
|
|
|
|
|
|
fn display_lines(&self) -> Vec<Line<'static>> {
|
|
|
|
|
|
self.lines.clone()
|
2025-07-30 10:05:40 -07:00
|
|
|
|
}
|
|
|
|
|
|
}
|
|
|
|
|
|
|
2025-08-11 12:40:12 -07:00
|
|
|
|
pub(crate) struct ExecCell {
|
|
|
|
|
|
pub(crate) command: Vec<String>,
|
|
|
|
|
|
pub(crate) parsed: Vec<ParsedCommand>,
|
|
|
|
|
|
pub(crate) output: Option<CommandOutput>,
|
|
|
|
|
|
}
|
2025-08-14 14:10:05 -04:00
|
|
|
|
impl HistoryCell for ExecCell {
|
|
|
|
|
|
fn display_lines(&self) -> Vec<Line<'static>> {
|
|
|
|
|
|
exec_command_lines(&self.command, &self.parsed, self.output.as_ref())
|
|
|
|
|
|
}
|
|
|
|
|
|
}
|
2025-08-11 12:40:12 -07:00
|
|
|
|
|
2025-08-14 14:10:05 -04:00
|
|
|
|
impl WidgetRef for &ExecCell {
|
|
|
|
|
|
fn render_ref(&self, area: Rect, buf: &mut Buffer) {
|
|
|
|
|
|
Paragraph::new(Text::from(self.display_lines()))
|
|
|
|
|
|
.wrap(Wrap { trim: false })
|
|
|
|
|
|
.render(area, buf);
|
|
|
|
|
|
}
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
struct CompletedMcpToolCallWithImageOutput {
|
|
|
|
|
|
_image: DynamicImage,
|
|
|
|
|
|
}
|
|
|
|
|
|
impl HistoryCell for CompletedMcpToolCallWithImageOutput {
|
|
|
|
|
|
fn display_lines(&self) -> Vec<Line<'static>> {
|
|
|
|
|
|
vec![
|
|
|
|
|
|
Line::from("tool result (image output omitted)"),
|
|
|
|
|
|
Line::from(""),
|
|
|
|
|
|
]
|
|
|
|
|
|
}
|
feat: initial import of Rust implementation of Codex CLI in codex-rs/ (#629)
As stated in `codex-rs/README.md`:
Today, Codex CLI is written in TypeScript and requires Node.js 22+ to
run it. For a number of users, this runtime requirement inhibits
adoption: they would be better served by a standalone executable. As
maintainers, we want Codex to run efficiently in a wide range of
environments with minimal overhead. We also want to take advantage of
operating system-specific APIs to provide better sandboxing, where
possible.
To that end, we are moving forward with a Rust implementation of Codex
CLI contained in this folder, which has the following benefits:
- The CLI compiles to small, standalone, platform-specific binaries.
- Can make direct, native calls to
[seccomp](https://man7.org/linux/man-pages/man2/seccomp.2.html) and
[landlock](https://man7.org/linux/man-pages/man7/landlock.7.html) in
order to support sandboxing on Linux.
- No runtime garbage collection, resulting in lower memory consumption
and better, more predictable performance.
Currently, the Rust implementation is materially behind the TypeScript
implementation in functionality, so continue to use the TypeScript
implmentation for the time being. We will publish native executables via
GitHub Releases as soon as we feel the Rust version is usable.
2025-04-24 13:31:40 -07:00
|
|
|
|
}
|
|
|
|
|
|
|
2025-08-10 21:32:56 -07:00
|
|
|
|
const TOOL_CALL_MAX_LINES: usize = 5;
|
2025-05-06 16:12:15 -07:00
|
|
|
|
|
2025-08-07 04:02:58 -07:00
|
|
|
|
fn title_case(s: &str) -> String {
|
|
|
|
|
|
if s.is_empty() {
|
|
|
|
|
|
return String::new();
|
|
|
|
|
|
}
|
|
|
|
|
|
let mut chars = s.chars();
|
|
|
|
|
|
let first = match chars.next() {
|
|
|
|
|
|
Some(c) => c,
|
|
|
|
|
|
None => return String::new(),
|
|
|
|
|
|
};
|
|
|
|
|
|
let rest: String = chars.as_str().to_ascii_lowercase();
|
|
|
|
|
|
first.to_uppercase().collect::<String>() + &rest
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
fn pretty_provider_name(id: &str) -> String {
|
|
|
|
|
|
if id.eq_ignore_ascii_case("openai") {
|
|
|
|
|
|
"OpenAI".to_string()
|
|
|
|
|
|
} else {
|
|
|
|
|
|
title_case(id)
|
|
|
|
|
|
}
|
|
|
|
|
|
}
|
|
|
|
|
|
|
2025-08-14 14:10:05 -04:00
|
|
|
|
pub(crate) fn new_background_event(message: String) -> PlainHistoryCell {
|
|
|
|
|
|
let mut lines: Vec<Line<'static>> = Vec::new();
|
|
|
|
|
|
lines.push(Line::from("event".dim()));
|
|
|
|
|
|
lines.extend(message.lines().map(|line| ansi_escape_line(line).dim()));
|
|
|
|
|
|
lines.push(Line::from(""));
|
|
|
|
|
|
PlainHistoryCell { lines }
|
|
|
|
|
|
}
|
2025-08-06 12:03:45 -07:00
|
|
|
|
|
2025-08-14 14:10:05 -04:00
|
|
|
|
pub(crate) fn new_session_info(
|
|
|
|
|
|
config: &Config,
|
|
|
|
|
|
event: SessionConfiguredEvent,
|
|
|
|
|
|
is_first_event: bool,
|
|
|
|
|
|
) -> PlainHistoryCell {
|
|
|
|
|
|
let SessionConfiguredEvent {
|
|
|
|
|
|
model,
|
|
|
|
|
|
session_id: _,
|
|
|
|
|
|
history_log_id: _,
|
|
|
|
|
|
history_entry_count: _,
|
|
|
|
|
|
} = event;
|
|
|
|
|
|
if is_first_event {
|
|
|
|
|
|
let cwd_str = match relativize_to_home(&config.cwd) {
|
|
|
|
|
|
Some(rel) if !rel.as_os_str().is_empty() => format!("~/{}", rel.display()),
|
|
|
|
|
|
Some(_) => "~".to_string(),
|
|
|
|
|
|
None => config.cwd.display().to_string(),
|
|
|
|
|
|
};
|
2025-08-12 17:37:28 -07:00
|
|
|
|
|
2025-08-14 14:10:05 -04:00
|
|
|
|
let lines: Vec<Line<'static>> = vec![
|
|
|
|
|
|
Line::from(vec![
|
|
|
|
|
|
Span::raw(">_ ").dim(),
|
|
|
|
|
|
Span::styled(
|
|
|
|
|
|
"You are using OpenAI Codex in",
|
|
|
|
|
|
Style::default().add_modifier(Modifier::BOLD),
|
|
|
|
|
|
),
|
|
|
|
|
|
Span::raw(format!(" {cwd_str}")).dim(),
|
|
|
|
|
|
]),
|
|
|
|
|
|
Line::from("".dim()),
|
|
|
|
|
|
Line::from(" To get started, describe a task or try one of these commands:".dim()),
|
|
|
|
|
|
Line::from("".dim()),
|
|
|
|
|
|
Line::from(format!(" /init - {}", SlashCommand::Init.description()).dim()),
|
|
|
|
|
|
Line::from(format!(" /status - {}", SlashCommand::Status.description()).dim()),
|
|
|
|
|
|
Line::from(format!(" /diff - {}", SlashCommand::Diff.description()).dim()),
|
|
|
|
|
|
Line::from(format!(" /prompts - {}", SlashCommand::Prompts.description()).dim()),
|
|
|
|
|
|
Line::from("".dim()),
|
|
|
|
|
|
];
|
|
|
|
|
|
PlainHistoryCell { lines }
|
|
|
|
|
|
} else if config.model == model {
|
|
|
|
|
|
PlainHistoryCell { lines: Vec::new() }
|
|
|
|
|
|
} else {
|
|
|
|
|
|
let lines = vec![
|
|
|
|
|
|
Line::from("model changed:".magenta().bold()),
|
|
|
|
|
|
Line::from(format!("requested: {}", config.model)),
|
|
|
|
|
|
Line::from(format!("used: {model}")),
|
|
|
|
|
|
Line::from(""),
|
|
|
|
|
|
];
|
|
|
|
|
|
PlainHistoryCell { lines }
|
2025-08-06 12:03:45 -07:00
|
|
|
|
}
|
2025-08-14 14:10:05 -04:00
|
|
|
|
}
|
2025-08-06 12:03:45 -07:00
|
|
|
|
|
2025-08-14 14:10:05 -04:00
|
|
|
|
pub(crate) fn new_user_prompt(message: String) -> PlainHistoryCell {
|
|
|
|
|
|
let mut lines: Vec<Line<'static>> = Vec::new();
|
|
|
|
|
|
lines.push(Line::from("user".cyan().bold()));
|
|
|
|
|
|
lines.extend(message.lines().map(|l| Line::from(l.to_string())));
|
|
|
|
|
|
lines.push(Line::from(""));
|
2025-05-30 23:24:36 -07:00
|
|
|
|
|
2025-08-14 14:10:05 -04:00
|
|
|
|
PlainHistoryCell { lines }
|
|
|
|
|
|
}
|
2025-05-08 21:46:06 -07:00
|
|
|
|
|
2025-08-14 14:10:05 -04:00
|
|
|
|
pub(crate) fn new_active_exec_command(
|
|
|
|
|
|
command: Vec<String>,
|
|
|
|
|
|
parsed: Vec<ParsedCommand>,
|
|
|
|
|
|
) -> ExecCell {
|
|
|
|
|
|
new_exec_cell(command, parsed, None)
|
|
|
|
|
|
}
|
feat: initial import of Rust implementation of Codex CLI in codex-rs/ (#629)
As stated in `codex-rs/README.md`:
Today, Codex CLI is written in TypeScript and requires Node.js 22+ to
run it. For a number of users, this runtime requirement inhibits
adoption: they would be better served by a standalone executable. As
maintainers, we want Codex to run efficiently in a wide range of
environments with minimal overhead. We also want to take advantage of
operating system-specific APIs to provide better sandboxing, where
possible.
To that end, we are moving forward with a Rust implementation of Codex
CLI contained in this folder, which has the following benefits:
- The CLI compiles to small, standalone, platform-specific binaries.
- Can make direct, native calls to
[seccomp](https://man7.org/linux/man-pages/man2/seccomp.2.html) and
[landlock](https://man7.org/linux/man-pages/man7/landlock.7.html) in
order to support sandboxing on Linux.
- No runtime garbage collection, resulting in lower memory consumption
and better, more predictable performance.
Currently, the Rust implementation is materially behind the TypeScript
implementation in functionality, so continue to use the TypeScript
implmentation for the time being. We will publish native executables via
GitHub Releases as soon as we feel the Rust version is usable.
2025-04-24 13:31:40 -07:00
|
|
|
|
|
2025-08-14 14:10:05 -04:00
|
|
|
|
pub(crate) fn new_completed_exec_command(
|
|
|
|
|
|
command: Vec<String>,
|
|
|
|
|
|
parsed: Vec<ParsedCommand>,
|
|
|
|
|
|
output: CommandOutput,
|
|
|
|
|
|
) -> ExecCell {
|
|
|
|
|
|
new_exec_cell(command, parsed, Some(output))
|
|
|
|
|
|
}
|
feat: initial import of Rust implementation of Codex CLI in codex-rs/ (#629)
As stated in `codex-rs/README.md`:
Today, Codex CLI is written in TypeScript and requires Node.js 22+ to
run it. For a number of users, this runtime requirement inhibits
adoption: they would be better served by a standalone executable. As
maintainers, we want Codex to run efficiently in a wide range of
environments with minimal overhead. We also want to take advantage of
operating system-specific APIs to provide better sandboxing, where
possible.
To that end, we are moving forward with a Rust implementation of Codex
CLI contained in this folder, which has the following benefits:
- The CLI compiles to small, standalone, platform-specific binaries.
- Can make direct, native calls to
[seccomp](https://man7.org/linux/man-pages/man2/seccomp.2.html) and
[landlock](https://man7.org/linux/man-pages/man7/landlock.7.html) in
order to support sandboxing on Linux.
- No runtime garbage collection, resulting in lower memory consumption
and better, more predictable performance.
Currently, the Rust implementation is materially behind the TypeScript
implementation in functionality, so continue to use the TypeScript
implmentation for the time being. We will publish native executables via
GitHub Releases as soon as we feel the Rust version is usable.
2025-04-24 13:31:40 -07:00
|
|
|
|
|
2025-08-14 14:10:05 -04:00
|
|
|
|
fn new_exec_cell(
|
|
|
|
|
|
command: Vec<String>,
|
|
|
|
|
|
parsed: Vec<ParsedCommand>,
|
|
|
|
|
|
output: Option<CommandOutput>,
|
|
|
|
|
|
) -> ExecCell {
|
|
|
|
|
|
ExecCell {
|
|
|
|
|
|
command,
|
|
|
|
|
|
parsed,
|
|
|
|
|
|
output,
|
2025-08-11 11:26:15 -07:00
|
|
|
|
}
|
2025-08-14 14:10:05 -04:00
|
|
|
|
}
|
feat: initial import of Rust implementation of Codex CLI in codex-rs/ (#629)
As stated in `codex-rs/README.md`:
Today, Codex CLI is written in TypeScript and requires Node.js 22+ to
run it. For a number of users, this runtime requirement inhibits
adoption: they would be better served by a standalone executable. As
maintainers, we want Codex to run efficiently in a wide range of
environments with minimal overhead. We also want to take advantage of
operating system-specific APIs to provide better sandboxing, where
possible.
To that end, we are moving forward with a Rust implementation of Codex
CLI contained in this folder, which has the following benefits:
- The CLI compiles to small, standalone, platform-specific binaries.
- Can make direct, native calls to
[seccomp](https://man7.org/linux/man-pages/man2/seccomp.2.html) and
[landlock](https://man7.org/linux/man-pages/man7/landlock.7.html) in
order to support sandboxing on Linux.
- No runtime garbage collection, resulting in lower memory consumption
and better, more predictable performance.
Currently, the Rust implementation is materially behind the TypeScript
implementation in functionality, so continue to use the TypeScript
implmentation for the time being. We will publish native executables via
GitHub Releases as soon as we feel the Rust version is usable.
2025-04-24 13:31:40 -07:00
|
|
|
|
|
2025-08-14 14:10:05 -04:00
|
|
|
|
fn exec_command_lines(
|
|
|
|
|
|
command: &[String],
|
|
|
|
|
|
parsed: &[ParsedCommand],
|
|
|
|
|
|
output: Option<&CommandOutput>,
|
|
|
|
|
|
) -> Vec<Line<'static>> {
|
|
|
|
|
|
match parsed.is_empty() {
|
|
|
|
|
|
true => new_exec_command_generic(command, output),
|
|
|
|
|
|
false => new_parsed_command(parsed, output),
|
2025-08-11 11:43:58 -07:00
|
|
|
|
}
|
2025-08-14 14:10:05 -04:00
|
|
|
|
}
|
2025-08-11 11:43:58 -07:00
|
|
|
|
|
2025-08-14 14:10:05 -04:00
|
|
|
|
fn new_parsed_command(
|
|
|
|
|
|
parsed_commands: &[ParsedCommand],
|
|
|
|
|
|
output: Option<&CommandOutput>,
|
|
|
|
|
|
) -> Vec<Line<'static>> {
|
|
|
|
|
|
let mut lines: Vec<Line> = vec![match output {
|
|
|
|
|
|
None => Line::from("⚙︎ Working".magenta().bold()),
|
|
|
|
|
|
Some(o) if o.exit_code == 0 => Line::from("✓ Completed".green().bold()),
|
|
|
|
|
|
Some(o) => Line::from(format!("✗ Failed (exit {})", o.exit_code).red().bold()),
|
|
|
|
|
|
}];
|
|
|
|
|
|
|
|
|
|
|
|
for (i, parsed) in parsed_commands.iter().enumerate() {
|
|
|
|
|
|
let text = match parsed {
|
|
|
|
|
|
ParsedCommand::Read { name, .. } => format!("📖 {name}"),
|
|
|
|
|
|
ParsedCommand::ListFiles { cmd, path } => match path {
|
|
|
|
|
|
Some(p) => format!("📂 {p}"),
|
|
|
|
|
|
None => format!("📂 {}", shlex_join_safe(cmd)),
|
|
|
|
|
|
},
|
|
|
|
|
|
ParsedCommand::Search { query, path, cmd } => match (query, path) {
|
|
|
|
|
|
(Some(q), Some(p)) => format!("🔎 {q} in {p}"),
|
|
|
|
|
|
(Some(q), None) => format!("🔎 {q}"),
|
|
|
|
|
|
(None, Some(p)) => format!("🔎 {p}"),
|
|
|
|
|
|
(None, None) => format!("🔎 {}", shlex_join_safe(cmd)),
|
|
|
|
|
|
},
|
|
|
|
|
|
ParsedCommand::Format { .. } => "✨ Formatting".to_string(),
|
|
|
|
|
|
ParsedCommand::Test { cmd } => format!("🧪 {}", shlex_join_safe(cmd)),
|
|
|
|
|
|
ParsedCommand::Lint { cmd, .. } => format!("🧹 {}", shlex_join_safe(cmd)),
|
|
|
|
|
|
ParsedCommand::Unknown { cmd } => format!("⌨️ {}", shlex_join_safe(cmd)),
|
|
|
|
|
|
};
|
2025-08-11 11:26:15 -07:00
|
|
|
|
|
2025-08-14 14:10:05 -04:00
|
|
|
|
let first_prefix = if i == 0 { " └ " } else { " " };
|
|
|
|
|
|
for (j, line_text) in text.lines().enumerate() {
|
|
|
|
|
|
let prefix = if j == 0 { first_prefix } else { " " };
|
|
|
|
|
|
lines.push(Line::from(vec![
|
|
|
|
|
|
Span::styled(prefix, Style::default().add_modifier(Modifier::DIM)),
|
|
|
|
|
|
Span::styled(line_text.to_string(), Style::default().fg(LIGHT_BLUE)),
|
|
|
|
|
|
]));
|
2025-08-08 10:52:24 -07:00
|
|
|
|
}
|
2025-08-11 11:26:15 -07:00
|
|
|
|
}
|
|
|
|
|
|
|
2025-08-14 14:10:05 -04:00
|
|
|
|
lines.extend(output_lines(output, true, false));
|
|
|
|
|
|
lines.push(Line::from(""));
|
2025-08-11 11:26:15 -07:00
|
|
|
|
|
2025-08-14 14:10:05 -04:00
|
|
|
|
lines
|
|
|
|
|
|
}
|
2025-08-11 11:26:15 -07:00
|
|
|
|
|
2025-08-14 14:10:05 -04:00
|
|
|
|
fn new_exec_command_generic(
|
|
|
|
|
|
command: &[String],
|
|
|
|
|
|
output: Option<&CommandOutput>,
|
|
|
|
|
|
) -> Vec<Line<'static>> {
|
|
|
|
|
|
let mut lines: Vec<Line<'static>> = Vec::new();
|
|
|
|
|
|
let command_escaped = strip_bash_lc_and_escape(command);
|
|
|
|
|
|
let mut cmd_lines = command_escaped.lines();
|
|
|
|
|
|
if let Some(first) = cmd_lines.next() {
|
|
|
|
|
|
lines.push(Line::from(vec![
|
|
|
|
|
|
"⚡ Running ".to_string().magenta(),
|
|
|
|
|
|
first.to_string().into(),
|
|
|
|
|
|
]));
|
|
|
|
|
|
} else {
|
|
|
|
|
|
lines.push(Line::from("⚡ Running".to_string().magenta()));
|
|
|
|
|
|
}
|
|
|
|
|
|
for cont in cmd_lines {
|
|
|
|
|
|
lines.push(Line::from(cont.to_string()));
|
|
|
|
|
|
}
|
2025-08-11 11:26:15 -07:00
|
|
|
|
|
2025-08-14 14:10:05 -04:00
|
|
|
|
lines.extend(output_lines(output, false, true));
|
feat: initial import of Rust implementation of Codex CLI in codex-rs/ (#629)
As stated in `codex-rs/README.md`:
Today, Codex CLI is written in TypeScript and requires Node.js 22+ to
run it. For a number of users, this runtime requirement inhibits
adoption: they would be better served by a standalone executable. As
maintainers, we want Codex to run efficiently in a wide range of
environments with minimal overhead. We also want to take advantage of
operating system-specific APIs to provide better sandboxing, where
possible.
To that end, we are moving forward with a Rust implementation of Codex
CLI contained in this folder, which has the following benefits:
- The CLI compiles to small, standalone, platform-specific binaries.
- Can make direct, native calls to
[seccomp](https://man7.org/linux/man-pages/man2/seccomp.2.html) and
[landlock](https://man7.org/linux/man-pages/man7/landlock.7.html) in
order to support sandboxing on Linux.
- No runtime garbage collection, resulting in lower memory consumption
and better, more predictable performance.
Currently, the Rust implementation is materially behind the TypeScript
implementation in functionality, so continue to use the TypeScript
implmentation for the time being. We will publish native executables via
GitHub Releases as soon as we feel the Rust version is usable.
2025-04-24 13:31:40 -07:00
|
|
|
|
|
2025-08-14 14:10:05 -04:00
|
|
|
|
lines
|
|
|
|
|
|
}
|
feat: initial import of Rust implementation of Codex CLI in codex-rs/ (#629)
As stated in `codex-rs/README.md`:
Today, Codex CLI is written in TypeScript and requires Node.js 22+ to
run it. For a number of users, this runtime requirement inhibits
adoption: they would be better served by a standalone executable. As
maintainers, we want Codex to run efficiently in a wide range of
environments with minimal overhead. We also want to take advantage of
operating system-specific APIs to provide better sandboxing, where
possible.
To that end, we are moving forward with a Rust implementation of Codex
CLI contained in this folder, which has the following benefits:
- The CLI compiles to small, standalone, platform-specific binaries.
- Can make direct, native calls to
[seccomp](https://man7.org/linux/man-pages/man2/seccomp.2.html) and
[landlock](https://man7.org/linux/man-pages/man7/landlock.7.html) in
order to support sandboxing on Linux.
- No runtime garbage collection, resulting in lower memory consumption
and better, more predictable performance.
Currently, the Rust implementation is materially behind the TypeScript
implementation in functionality, so continue to use the TypeScript
implmentation for the time being. We will publish native executables via
GitHub Releases as soon as we feel the Rust version is usable.
2025-04-24 13:31:40 -07:00
|
|
|
|
|
2025-08-14 14:10:05 -04:00
|
|
|
|
pub(crate) fn new_active_mcp_tool_call(invocation: McpInvocation) -> PlainHistoryCell {
|
|
|
|
|
|
let title_line = Line::from(vec!["tool".magenta(), " running...".dim()]);
|
|
|
|
|
|
let lines: Vec<Line> = vec![
|
|
|
|
|
|
title_line,
|
|
|
|
|
|
format_mcp_invocation(invocation.clone()),
|
|
|
|
|
|
Line::from(""),
|
|
|
|
|
|
];
|
feat: initial import of Rust implementation of Codex CLI in codex-rs/ (#629)
As stated in `codex-rs/README.md`:
Today, Codex CLI is written in TypeScript and requires Node.js 22+ to
run it. For a number of users, this runtime requirement inhibits
adoption: they would be better served by a standalone executable. As
maintainers, we want Codex to run efficiently in a wide range of
environments with minimal overhead. We also want to take advantage of
operating system-specific APIs to provide better sandboxing, where
possible.
To that end, we are moving forward with a Rust implementation of Codex
CLI contained in this folder, which has the following benefits:
- The CLI compiles to small, standalone, platform-specific binaries.
- Can make direct, native calls to
[seccomp](https://man7.org/linux/man-pages/man2/seccomp.2.html) and
[landlock](https://man7.org/linux/man-pages/man7/landlock.7.html) in
order to support sandboxing on Linux.
- No runtime garbage collection, resulting in lower memory consumption
and better, more predictable performance.
Currently, the Rust implementation is materially behind the TypeScript
implementation in functionality, so continue to use the TypeScript
implmentation for the time being. We will publish native executables via
GitHub Releases as soon as we feel the Rust version is usable.
2025-04-24 13:31:40 -07:00
|
|
|
|
|
2025-08-14 14:10:05 -04:00
|
|
|
|
PlainHistoryCell { lines }
|
|
|
|
|
|
}
|
feat: initial import of Rust implementation of Codex CLI in codex-rs/ (#629)
As stated in `codex-rs/README.md`:
Today, Codex CLI is written in TypeScript and requires Node.js 22+ to
run it. For a number of users, this runtime requirement inhibits
adoption: they would be better served by a standalone executable. As
maintainers, we want Codex to run efficiently in a wide range of
environments with minimal overhead. We also want to take advantage of
operating system-specific APIs to provide better sandboxing, where
possible.
To that end, we are moving forward with a Rust implementation of Codex
CLI contained in this folder, which has the following benefits:
- The CLI compiles to small, standalone, platform-specific binaries.
- Can make direct, native calls to
[seccomp](https://man7.org/linux/man-pages/man2/seccomp.2.html) and
[landlock](https://man7.org/linux/man-pages/man7/landlock.7.html) in
order to support sandboxing on Linux.
- No runtime garbage collection, resulting in lower memory consumption
and better, more predictable performance.
Currently, the Rust implementation is materially behind the TypeScript
implementation in functionality, so continue to use the TypeScript
implmentation for the time being. We will publish native executables via
GitHub Releases as soon as we feel the Rust version is usable.
2025-04-24 13:31:40 -07:00
|
|
|
|
|
2025-08-14 14:10:05 -04:00
|
|
|
|
/// If the first content is an image, return a new cell with the image.
|
|
|
|
|
|
/// TODO(rgwood-dd): Handle images properly even if they're not the first result.
|
|
|
|
|
|
fn try_new_completed_mcp_tool_call_with_image_output(
|
|
|
|
|
|
result: &Result<mcp_types::CallToolResult, String>,
|
|
|
|
|
|
) -> Option<CompletedMcpToolCallWithImageOutput> {
|
|
|
|
|
|
match result {
|
|
|
|
|
|
Ok(mcp_types::CallToolResult { content, .. }) => {
|
|
|
|
|
|
if let Some(mcp_types::ContentBlock::ImageContent(image)) = content.first() {
|
|
|
|
|
|
let raw_data = match base64::engine::general_purpose::STANDARD.decode(&image.data) {
|
|
|
|
|
|
Ok(data) => data,
|
|
|
|
|
|
Err(e) => {
|
|
|
|
|
|
error!("Failed to decode image data: {e}");
|
|
|
|
|
|
return None;
|
|
|
|
|
|
}
|
|
|
|
|
|
};
|
|
|
|
|
|
let reader = match ImageReader::new(Cursor::new(raw_data)).with_guessed_format() {
|
|
|
|
|
|
Ok(reader) => reader,
|
|
|
|
|
|
Err(e) => {
|
|
|
|
|
|
error!("Failed to guess image format: {e}");
|
|
|
|
|
|
return None;
|
|
|
|
|
|
}
|
|
|
|
|
|
};
|
feat: initial import of Rust implementation of Codex CLI in codex-rs/ (#629)
As stated in `codex-rs/README.md`:
Today, Codex CLI is written in TypeScript and requires Node.js 22+ to
run it. For a number of users, this runtime requirement inhibits
adoption: they would be better served by a standalone executable. As
maintainers, we want Codex to run efficiently in a wide range of
environments with minimal overhead. We also want to take advantage of
operating system-specific APIs to provide better sandboxing, where
possible.
To that end, we are moving forward with a Rust implementation of Codex
CLI contained in this folder, which has the following benefits:
- The CLI compiles to small, standalone, platform-specific binaries.
- Can make direct, native calls to
[seccomp](https://man7.org/linux/man-pages/man2/seccomp.2.html) and
[landlock](https://man7.org/linux/man-pages/man7/landlock.7.html) in
order to support sandboxing on Linux.
- No runtime garbage collection, resulting in lower memory consumption
and better, more predictable performance.
Currently, the Rust implementation is materially behind the TypeScript
implementation in functionality, so continue to use the TypeScript
implmentation for the time being. We will publish native executables via
GitHub Releases as soon as we feel the Rust version is usable.
2025-04-24 13:31:40 -07:00
|
|
|
|
|
2025-08-14 14:10:05 -04:00
|
|
|
|
let image = match reader.decode() {
|
|
|
|
|
|
Ok(image) => image,
|
|
|
|
|
|
Err(e) => {
|
|
|
|
|
|
error!("Image decoding failed: {e}");
|
|
|
|
|
|
return None;
|
|
|
|
|
|
}
|
|
|
|
|
|
};
|
2025-05-06 16:12:15 -07:00
|
|
|
|
|
2025-08-14 14:10:05 -04:00
|
|
|
|
Some(CompletedMcpToolCallWithImageOutput { _image: image })
|
|
|
|
|
|
} else {
|
|
|
|
|
|
None
|
|
|
|
|
|
}
|
2025-05-06 16:12:15 -07:00
|
|
|
|
}
|
2025-08-14 14:10:05 -04:00
|
|
|
|
_ => None,
|
2025-05-06 16:12:15 -07:00
|
|
|
|
}
|
2025-08-14 14:10:05 -04:00
|
|
|
|
}
|
2025-05-06 16:12:15 -07:00
|
|
|
|
|
2025-08-14 14:10:05 -04:00
|
|
|
|
pub(crate) fn new_completed_mcp_tool_call(
|
|
|
|
|
|
num_cols: usize,
|
|
|
|
|
|
invocation: McpInvocation,
|
|
|
|
|
|
duration: Duration,
|
|
|
|
|
|
success: bool,
|
|
|
|
|
|
result: Result<mcp_types::CallToolResult, String>,
|
|
|
|
|
|
) -> Box<dyn HistoryCell> {
|
|
|
|
|
|
if let Some(cell) = try_new_completed_mcp_tool_call_with_image_output(&result) {
|
|
|
|
|
|
return Box::new(cell);
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
let duration = format_duration(duration);
|
|
|
|
|
|
let status_str = if success { "success" } else { "failed" };
|
|
|
|
|
|
let title_line = Line::from(vec![
|
|
|
|
|
|
"tool".magenta(),
|
|
|
|
|
|
" ".into(),
|
|
|
|
|
|
if success {
|
|
|
|
|
|
status_str.green()
|
|
|
|
|
|
} else {
|
|
|
|
|
|
status_str.red()
|
|
|
|
|
|
},
|
|
|
|
|
|
format!(", duration: {duration}").dim(),
|
|
|
|
|
|
]);
|
|
|
|
|
|
|
|
|
|
|
|
let mut lines: Vec<Line<'static>> = Vec::new();
|
|
|
|
|
|
lines.push(title_line);
|
|
|
|
|
|
lines.push(format_mcp_invocation(invocation));
|
|
|
|
|
|
|
|
|
|
|
|
match result {
|
|
|
|
|
|
Ok(mcp_types::CallToolResult { content, .. }) => {
|
|
|
|
|
|
if !content.is_empty() {
|
|
|
|
|
|
lines.push(Line::from(""));
|
fix: introduce ResponseInputItem::McpToolCallOutput variant (#1151)
The output of an MCP server tool call can be one of several types, but
to date, we treated all outputs as text by showing the serialized JSON
as the "tool output" in Codex:
https://github.com/openai/codex/blob/25a9949c49194d5a64de54a11bcc5b4724ac9bd5/codex-rs/mcp-types/src/lib.rs#L96-L101
This PR adds support for the `ImageContent` variant so we can now
display an image output from an MCP tool call.
In making this change, we introduce a new
`ResponseInputItem::McpToolCallOutput` variant so that we can work with
the `mcp_types::CallToolResult` directly when the function call is made
to an MCP server.
Though arguably the more significant change is the introduction of
`HistoryCell::CompletedMcpToolCallWithImageOutput`, which is a cell that
uses `ratatui_image` to render an image into the terminal. To support
this, we introduce `ImageRenderCache`, cache a
`ratatui_image::picker::Picker`, and `ensure_image_cache()` to cache the
appropriate scaled image data and dimensions based on the current
terminal size.
To test, I created a minimal `package.json`:
```json
{
"name": "kitty-mcp",
"version": "1.0.0",
"type": "module",
"description": "MCP that returns image of kitty",
"main": "index.js",
"dependencies": {
"@modelcontextprotocol/sdk": "^1.12.0"
}
}
```
with the following `index.js` to define the MCP server:
```js
#!/usr/bin/env node
import { McpServer } from "@modelcontextprotocol/sdk/server/mcp.js";
import { StdioServerTransport } from "@modelcontextprotocol/sdk/server/stdio.js";
import { readFile } from "node:fs/promises";
import { join } from "node:path";
const IMAGE_URI = "image://Ada.png";
const server = new McpServer({
name: "Demo",
version: "1.0.0",
});
server.tool(
"get-cat-image",
"If you need a cat image, this tool will provide one.",
async () => ({
content: [
{ type: "image", data: await getAdaPngBase64(), mimeType: "image/png" },
],
})
);
server.resource("Ada the Cat", IMAGE_URI, async (uri) => {
const base64Image = await getAdaPngBase64();
return {
contents: [
{
uri: uri.href,
mimeType: "image/png",
blob: base64Image,
},
],
};
});
async function getAdaPngBase64() {
const __dirname = new URL(".", import.meta.url).pathname;
// From https://github.com/benjajaja/ratatui-image/blob/9705ce2c59ec669abbce2924cbfd1f5ae22c9860/assets/Ada.png
const filePath = join(__dirname, "Ada.png");
const imageData = await readFile(filePath);
const base64Image = imageData.toString("base64");
return base64Image;
}
const transport = new StdioServerTransport();
await server.connect(transport);
```
With the local changes from this PR, I added the following to my
`config.toml`:
```toml
[mcp_servers.kitty]
command = "node"
args = ["/Users/mbolin/code/kitty-mcp/index.js"]
```
Running the TUI from source:
```
cargo run --bin codex -- --model o3 'I need a picture of a cat'
```
I get:
<img width="732" alt="image"
src="https://github.com/user-attachments/assets/bf80b721-9ca0-4d81-aec7-77d6899e2869"
/>
Now, that said, I have only tested in iTerm and there is definitely some
funny business with getting an accurate character-to-pixel ratio
(sometimes the `CompletedMcpToolCallWithImageOutput` thinks it needs 10
rows to render instead of 4), so there is still work to be done here.
2025-05-28 19:03:17 -07:00
|
|
|
|
|
2025-08-14 14:10:05 -04:00
|
|
|
|
for tool_call_result in content {
|
|
|
|
|
|
let line_text = match tool_call_result {
|
|
|
|
|
|
mcp_types::ContentBlock::TextContent(text) => {
|
|
|
|
|
|
format_and_truncate_tool_result(
|
|
|
|
|
|
&text.text,
|
|
|
|
|
|
TOOL_CALL_MAX_LINES,
|
|
|
|
|
|
num_cols,
|
|
|
|
|
|
)
|
|
|
|
|
|
}
|
|
|
|
|
|
mcp_types::ContentBlock::ImageContent(_) => {
|
|
|
|
|
|
// TODO show images even if they're not the first result, will require a refactor of `CompletedMcpToolCall`
|
|
|
|
|
|
"<image content>".to_string()
|
|
|
|
|
|
}
|
|
|
|
|
|
mcp_types::ContentBlock::AudioContent(_) => "<audio content>".to_string(),
|
|
|
|
|
|
mcp_types::ContentBlock::EmbeddedResource(resource) => {
|
|
|
|
|
|
let uri = match resource.resource {
|
|
|
|
|
|
EmbeddedResourceResource::TextResourceContents(text) => text.uri,
|
|
|
|
|
|
EmbeddedResourceResource::BlobResourceContents(blob) => blob.uri,
|
|
|
|
|
|
};
|
|
|
|
|
|
format!("embedded resource: {uri}")
|
|
|
|
|
|
}
|
|
|
|
|
|
mcp_types::ContentBlock::ResourceLink(ResourceLink { uri, .. }) => {
|
|
|
|
|
|
format!("link: {uri}")
|
fix: introduce ResponseInputItem::McpToolCallOutput variant (#1151)
The output of an MCP server tool call can be one of several types, but
to date, we treated all outputs as text by showing the serialized JSON
as the "tool output" in Codex:
https://github.com/openai/codex/blob/25a9949c49194d5a64de54a11bcc5b4724ac9bd5/codex-rs/mcp-types/src/lib.rs#L96-L101
This PR adds support for the `ImageContent` variant so we can now
display an image output from an MCP tool call.
In making this change, we introduce a new
`ResponseInputItem::McpToolCallOutput` variant so that we can work with
the `mcp_types::CallToolResult` directly when the function call is made
to an MCP server.
Though arguably the more significant change is the introduction of
`HistoryCell::CompletedMcpToolCallWithImageOutput`, which is a cell that
uses `ratatui_image` to render an image into the terminal. To support
this, we introduce `ImageRenderCache`, cache a
`ratatui_image::picker::Picker`, and `ensure_image_cache()` to cache the
appropriate scaled image data and dimensions based on the current
terminal size.
To test, I created a minimal `package.json`:
```json
{
"name": "kitty-mcp",
"version": "1.0.0",
"type": "module",
"description": "MCP that returns image of kitty",
"main": "index.js",
"dependencies": {
"@modelcontextprotocol/sdk": "^1.12.0"
}
}
```
with the following `index.js` to define the MCP server:
```js
#!/usr/bin/env node
import { McpServer } from "@modelcontextprotocol/sdk/server/mcp.js";
import { StdioServerTransport } from "@modelcontextprotocol/sdk/server/stdio.js";
import { readFile } from "node:fs/promises";
import { join } from "node:path";
const IMAGE_URI = "image://Ada.png";
const server = new McpServer({
name: "Demo",
version: "1.0.0",
});
server.tool(
"get-cat-image",
"If you need a cat image, this tool will provide one.",
async () => ({
content: [
{ type: "image", data: await getAdaPngBase64(), mimeType: "image/png" },
],
})
);
server.resource("Ada the Cat", IMAGE_URI, async (uri) => {
const base64Image = await getAdaPngBase64();
return {
contents: [
{
uri: uri.href,
mimeType: "image/png",
blob: base64Image,
},
],
};
});
async function getAdaPngBase64() {
const __dirname = new URL(".", import.meta.url).pathname;
// From https://github.com/benjajaja/ratatui-image/blob/9705ce2c59ec669abbce2924cbfd1f5ae22c9860/assets/Ada.png
const filePath = join(__dirname, "Ada.png");
const imageData = await readFile(filePath);
const base64Image = imageData.toString("base64");
return base64Image;
}
const transport = new StdioServerTransport();
await server.connect(transport);
```
With the local changes from this PR, I added the following to my
`config.toml`:
```toml
[mcp_servers.kitty]
command = "node"
args = ["/Users/mbolin/code/kitty-mcp/index.js"]
```
Running the TUI from source:
```
cargo run --bin codex -- --model o3 'I need a picture of a cat'
```
I get:
<img width="732" alt="image"
src="https://github.com/user-attachments/assets/bf80b721-9ca0-4d81-aec7-77d6899e2869"
/>
Now, that said, I have only tested in iTerm and there is definitely some
funny business with getting an accurate character-to-pixel ratio
(sometimes the `CompletedMcpToolCallWithImageOutput` thinks it needs 10
rows to render instead of 4), so there is still work to be done here.
2025-05-28 19:03:17 -07:00
|
|
|
|
}
|
|
|
|
|
|
};
|
2025-08-14 14:10:05 -04:00
|
|
|
|
lines.push(Line::styled(
|
|
|
|
|
|
line_text,
|
|
|
|
|
|
Style::default().add_modifier(Modifier::DIM),
|
|
|
|
|
|
));
|
fix: introduce ResponseInputItem::McpToolCallOutput variant (#1151)
The output of an MCP server tool call can be one of several types, but
to date, we treated all outputs as text by showing the serialized JSON
as the "tool output" in Codex:
https://github.com/openai/codex/blob/25a9949c49194d5a64de54a11bcc5b4724ac9bd5/codex-rs/mcp-types/src/lib.rs#L96-L101
This PR adds support for the `ImageContent` variant so we can now
display an image output from an MCP tool call.
In making this change, we introduce a new
`ResponseInputItem::McpToolCallOutput` variant so that we can work with
the `mcp_types::CallToolResult` directly when the function call is made
to an MCP server.
Though arguably the more significant change is the introduction of
`HistoryCell::CompletedMcpToolCallWithImageOutput`, which is a cell that
uses `ratatui_image` to render an image into the terminal. To support
this, we introduce `ImageRenderCache`, cache a
`ratatui_image::picker::Picker`, and `ensure_image_cache()` to cache the
appropriate scaled image data and dimensions based on the current
terminal size.
To test, I created a minimal `package.json`:
```json
{
"name": "kitty-mcp",
"version": "1.0.0",
"type": "module",
"description": "MCP that returns image of kitty",
"main": "index.js",
"dependencies": {
"@modelcontextprotocol/sdk": "^1.12.0"
}
}
```
with the following `index.js` to define the MCP server:
```js
#!/usr/bin/env node
import { McpServer } from "@modelcontextprotocol/sdk/server/mcp.js";
import { StdioServerTransport } from "@modelcontextprotocol/sdk/server/stdio.js";
import { readFile } from "node:fs/promises";
import { join } from "node:path";
const IMAGE_URI = "image://Ada.png";
const server = new McpServer({
name: "Demo",
version: "1.0.0",
});
server.tool(
"get-cat-image",
"If you need a cat image, this tool will provide one.",
async () => ({
content: [
{ type: "image", data: await getAdaPngBase64(), mimeType: "image/png" },
],
})
);
server.resource("Ada the Cat", IMAGE_URI, async (uri) => {
const base64Image = await getAdaPngBase64();
return {
contents: [
{
uri: uri.href,
mimeType: "image/png",
blob: base64Image,
},
],
};
});
async function getAdaPngBase64() {
const __dirname = new URL(".", import.meta.url).pathname;
// From https://github.com/benjajaja/ratatui-image/blob/9705ce2c59ec669abbce2924cbfd1f5ae22c9860/assets/Ada.png
const filePath = join(__dirname, "Ada.png");
const imageData = await readFile(filePath);
const base64Image = imageData.toString("base64");
return base64Image;
}
const transport = new StdioServerTransport();
await server.connect(transport);
```
With the local changes from this PR, I added the following to my
`config.toml`:
```toml
[mcp_servers.kitty]
command = "node"
args = ["/Users/mbolin/code/kitty-mcp/index.js"]
```
Running the TUI from source:
```
cargo run --bin codex -- --model o3 'I need a picture of a cat'
```
I get:
<img width="732" alt="image"
src="https://github.com/user-attachments/assets/bf80b721-9ca0-4d81-aec7-77d6899e2869"
/>
Now, that said, I have only tested in iTerm and there is definitely some
funny business with getting an accurate character-to-pixel ratio
(sometimes the `CompletedMcpToolCallWithImageOutput` thinks it needs 10
rows to render instead of 4), so there is still work to be done here.
2025-05-28 19:03:17 -07:00
|
|
|
|
}
|
|
|
|
|
|
}
|
|
|
|
|
|
|
2025-08-14 14:10:05 -04:00
|
|
|
|
lines.push(Line::from(""));
|
fix: introduce ResponseInputItem::McpToolCallOutput variant (#1151)
The output of an MCP server tool call can be one of several types, but
to date, we treated all outputs as text by showing the serialized JSON
as the "tool output" in Codex:
https://github.com/openai/codex/blob/25a9949c49194d5a64de54a11bcc5b4724ac9bd5/codex-rs/mcp-types/src/lib.rs#L96-L101
This PR adds support for the `ImageContent` variant so we can now
display an image output from an MCP tool call.
In making this change, we introduce a new
`ResponseInputItem::McpToolCallOutput` variant so that we can work with
the `mcp_types::CallToolResult` directly when the function call is made
to an MCP server.
Though arguably the more significant change is the introduction of
`HistoryCell::CompletedMcpToolCallWithImageOutput`, which is a cell that
uses `ratatui_image` to render an image into the terminal. To support
this, we introduce `ImageRenderCache`, cache a
`ratatui_image::picker::Picker`, and `ensure_image_cache()` to cache the
appropriate scaled image data and dimensions based on the current
terminal size.
To test, I created a minimal `package.json`:
```json
{
"name": "kitty-mcp",
"version": "1.0.0",
"type": "module",
"description": "MCP that returns image of kitty",
"main": "index.js",
"dependencies": {
"@modelcontextprotocol/sdk": "^1.12.0"
}
}
```
with the following `index.js` to define the MCP server:
```js
#!/usr/bin/env node
import { McpServer } from "@modelcontextprotocol/sdk/server/mcp.js";
import { StdioServerTransport } from "@modelcontextprotocol/sdk/server/stdio.js";
import { readFile } from "node:fs/promises";
import { join } from "node:path";
const IMAGE_URI = "image://Ada.png";
const server = new McpServer({
name: "Demo",
version: "1.0.0",
});
server.tool(
"get-cat-image",
"If you need a cat image, this tool will provide one.",
async () => ({
content: [
{ type: "image", data: await getAdaPngBase64(), mimeType: "image/png" },
],
})
);
server.resource("Ada the Cat", IMAGE_URI, async (uri) => {
const base64Image = await getAdaPngBase64();
return {
contents: [
{
uri: uri.href,
mimeType: "image/png",
blob: base64Image,
},
],
};
});
async function getAdaPngBase64() {
const __dirname = new URL(".", import.meta.url).pathname;
// From https://github.com/benjajaja/ratatui-image/blob/9705ce2c59ec669abbce2924cbfd1f5ae22c9860/assets/Ada.png
const filePath = join(__dirname, "Ada.png");
const imageData = await readFile(filePath);
const base64Image = imageData.toString("base64");
return base64Image;
}
const transport = new StdioServerTransport();
await server.connect(transport);
```
With the local changes from this PR, I added the following to my
`config.toml`:
```toml
[mcp_servers.kitty]
command = "node"
args = ["/Users/mbolin/code/kitty-mcp/index.js"]
```
Running the TUI from source:
```
cargo run --bin codex -- --model o3 'I need a picture of a cat'
```
I get:
<img width="732" alt="image"
src="https://github.com/user-attachments/assets/bf80b721-9ca0-4d81-aec7-77d6899e2869"
/>
Now, that said, I have only tested in iTerm and there is definitely some
funny business with getting an accurate character-to-pixel ratio
(sometimes the `CompletedMcpToolCallWithImageOutput` thinks it needs 10
rows to render instead of 4), so there is still work to be done here.
2025-05-28 19:03:17 -07:00
|
|
|
|
}
|
2025-08-14 14:10:05 -04:00
|
|
|
|
Err(e) => {
|
|
|
|
|
|
lines.push(Line::from(vec![
|
|
|
|
|
|
Span::styled(
|
|
|
|
|
|
"Error: ",
|
|
|
|
|
|
Style::default().fg(Color::Red).add_modifier(Modifier::BOLD),
|
|
|
|
|
|
),
|
|
|
|
|
|
Span::raw(e),
|
|
|
|
|
|
]));
|
2025-05-28 14:03:19 -07:00
|
|
|
|
}
|
2025-08-14 14:10:05 -04:00
|
|
|
|
};
|
feat: initial import of Rust implementation of Codex CLI in codex-rs/ (#629)
As stated in `codex-rs/README.md`:
Today, Codex CLI is written in TypeScript and requires Node.js 22+ to
run it. For a number of users, this runtime requirement inhibits
adoption: they would be better served by a standalone executable. As
maintainers, we want Codex to run efficiently in a wide range of
environments with minimal overhead. We also want to take advantage of
operating system-specific APIs to provide better sandboxing, where
possible.
To that end, we are moving forward with a Rust implementation of Codex
CLI contained in this folder, which has the following benefits:
- The CLI compiles to small, standalone, platform-specific binaries.
- Can make direct, native calls to
[seccomp](https://man7.org/linux/man-pages/man2/seccomp.2.html) and
[landlock](https://man7.org/linux/man-pages/man7/landlock.7.html) in
order to support sandboxing on Linux.
- No runtime garbage collection, resulting in lower memory consumption
and better, more predictable performance.
Currently, the Rust implementation is materially behind the TypeScript
implementation in functionality, so continue to use the TypeScript
implmentation for the time being. We will publish native executables via
GitHub Releases as soon as we feel the Rust version is usable.
2025-04-24 13:31:40 -07:00
|
|
|
|
|
2025-08-14 14:10:05 -04:00
|
|
|
|
Box::new(PlainHistoryCell { lines })
|
|
|
|
|
|
}
|
2025-06-26 13:03:31 -07:00
|
|
|
|
|
2025-08-14 14:10:05 -04:00
|
|
|
|
pub(crate) fn new_diff_output(message: String) -> PlainHistoryCell {
|
|
|
|
|
|
let mut lines: Vec<Line<'static>> = Vec::new();
|
|
|
|
|
|
lines.push(Line::from("/diff".magenta()));
|
2025-06-26 13:03:31 -07:00
|
|
|
|
|
2025-08-14 14:10:05 -04:00
|
|
|
|
if message.trim().is_empty() {
|
|
|
|
|
|
lines.push(Line::from("No changes detected.".italic()));
|
|
|
|
|
|
} else {
|
|
|
|
|
|
lines.extend(message.lines().map(ansi_escape_line));
|
2025-06-26 13:03:31 -07:00
|
|
|
|
}
|
|
|
|
|
|
|
2025-08-14 14:10:05 -04:00
|
|
|
|
lines.push(Line::from(""));
|
|
|
|
|
|
PlainHistoryCell { lines }
|
|
|
|
|
|
}
|
2025-08-07 04:02:58 -07:00
|
|
|
|
|
2025-08-14 14:10:05 -04:00
|
|
|
|
pub(crate) fn new_status_output(config: &Config, usage: &TokenUsage) -> PlainHistoryCell {
|
|
|
|
|
|
let mut lines: Vec<Line<'static>> = Vec::new();
|
|
|
|
|
|
lines.push(Line::from("/status".magenta()));
|
2025-08-05 23:57:52 -07:00
|
|
|
|
|
2025-08-14 14:10:05 -04:00
|
|
|
|
let config_entries = create_config_summary_entries(config);
|
|
|
|
|
|
let lookup = |k: &str| -> String {
|
|
|
|
|
|
config_entries
|
|
|
|
|
|
.iter()
|
|
|
|
|
|
.find(|(key, _)| *key == k)
|
|
|
|
|
|
.map(|(_, v)| v.clone())
|
|
|
|
|
|
.unwrap_or_default()
|
|
|
|
|
|
};
|
2025-08-07 01:27:45 -07:00
|
|
|
|
|
2025-08-14 14:10:05 -04:00
|
|
|
|
// 📂 Workspace
|
|
|
|
|
|
lines.push(Line::from(vec!["📂 ".into(), "Workspace".bold()]));
|
|
|
|
|
|
// Path (home-relative, e.g., ~/code/project)
|
|
|
|
|
|
let cwd_str = match relativize_to_home(&config.cwd) {
|
|
|
|
|
|
Some(rel) if !rel.as_os_str().is_empty() => format!("~/{}", rel.display()),
|
|
|
|
|
|
Some(_) => "~".to_string(),
|
|
|
|
|
|
None => config.cwd.display().to_string(),
|
|
|
|
|
|
};
|
|
|
|
|
|
lines.push(Line::from(vec![" • Path: ".into(), cwd_str.into()]));
|
|
|
|
|
|
// Approval mode (as-is)
|
|
|
|
|
|
lines.push(Line::from(vec![
|
|
|
|
|
|
" • Approval Mode: ".into(),
|
|
|
|
|
|
lookup("approval").into(),
|
|
|
|
|
|
]));
|
|
|
|
|
|
// Sandbox (simplified name only)
|
|
|
|
|
|
let sandbox_name = match &config.sandbox_policy {
|
|
|
|
|
|
SandboxPolicy::DangerFullAccess => "danger-full-access",
|
|
|
|
|
|
SandboxPolicy::ReadOnly => "read-only",
|
|
|
|
|
|
SandboxPolicy::WorkspaceWrite { .. } => "workspace-write",
|
|
|
|
|
|
};
|
|
|
|
|
|
lines.push(Line::from(vec![
|
|
|
|
|
|
" • Sandbox: ".into(),
|
|
|
|
|
|
sandbox_name.into(),
|
|
|
|
|
|
]));
|
|
|
|
|
|
|
|
|
|
|
|
lines.push(Line::from(""));
|
|
|
|
|
|
|
|
|
|
|
|
// 👤 Account (only if ChatGPT tokens exist), shown under the first block
|
|
|
|
|
|
let auth_file = get_auth_file(&config.codex_home);
|
|
|
|
|
|
if let Ok(auth) = try_read_auth_json(&auth_file) {
|
|
|
|
|
|
if let Some(tokens) = auth.tokens.clone() {
|
|
|
|
|
|
lines.push(Line::from(vec!["👤 ".into(), "Account".bold()]));
|
|
|
|
|
|
lines.push(Line::from(" • Signed in with ChatGPT"));
|
|
|
|
|
|
|
|
|
|
|
|
let info = tokens.id_token;
|
|
|
|
|
|
if let Some(email) = &info.email {
|
|
|
|
|
|
lines.push(Line::from(vec![" • Login: ".into(), email.clone().into()]));
|
|
|
|
|
|
}
|
2025-08-07 01:27:45 -07:00
|
|
|
|
|
2025-08-14 14:10:05 -04:00
|
|
|
|
match auth.openai_api_key.as_deref() {
|
|
|
|
|
|
Some(key) if !key.is_empty() => {
|
|
|
|
|
|
lines.push(Line::from(
|
|
|
|
|
|
" • Using API key. Run codex login to use ChatGPT plan",
|
|
|
|
|
|
));
|
2025-08-07 04:02:58 -07:00
|
|
|
|
}
|
2025-08-14 14:10:05 -04:00
|
|
|
|
_ => {
|
|
|
|
|
|
let plan_text = info
|
|
|
|
|
|
.get_chatgpt_plan_type()
|
|
|
|
|
|
.map(|s| title_case(&s))
|
|
|
|
|
|
.unwrap_or_else(|| "Unknown".to_string());
|
|
|
|
|
|
lines.push(Line::from(vec![" • Plan: ".into(), plan_text.into()]));
|
2025-08-07 01:27:45 -07:00
|
|
|
|
}
|
|
|
|
|
|
}
|
|
|
|
|
|
|
2025-08-14 14:10:05 -04:00
|
|
|
|
lines.push(Line::from(""));
|
2025-08-07 04:02:58 -07:00
|
|
|
|
}
|
2025-08-14 14:10:05 -04:00
|
|
|
|
}
|
2025-08-07 04:02:58 -07:00
|
|
|
|
|
2025-08-14 14:10:05 -04:00
|
|
|
|
// 🧠 Model
|
|
|
|
|
|
lines.push(Line::from(vec!["🧠 ".into(), "Model".bold()]));
|
|
|
|
|
|
lines.push(Line::from(vec![
|
|
|
|
|
|
" • Name: ".into(),
|
|
|
|
|
|
config.model.clone().into(),
|
|
|
|
|
|
]));
|
|
|
|
|
|
let provider_disp = pretty_provider_name(&config.model_provider_id);
|
|
|
|
|
|
lines.push(Line::from(vec![
|
|
|
|
|
|
" • Provider: ".into(),
|
|
|
|
|
|
provider_disp.into(),
|
|
|
|
|
|
]));
|
|
|
|
|
|
// Only show Reasoning fields if present in config summary
|
|
|
|
|
|
let reff = lookup("reasoning effort");
|
|
|
|
|
|
if !reff.is_empty() {
|
2025-08-05 23:57:52 -07:00
|
|
|
|
lines.push(Line::from(vec![
|
2025-08-14 14:10:05 -04:00
|
|
|
|
" • Reasoning Effort: ".into(),
|
|
|
|
|
|
title_case(&reff).into(),
|
2025-08-05 23:57:52 -07:00
|
|
|
|
]));
|
2025-08-14 14:10:05 -04:00
|
|
|
|
}
|
|
|
|
|
|
let rsum = lookup("reasoning summaries");
|
|
|
|
|
|
if !rsum.is_empty() {
|
2025-08-05 23:57:52 -07:00
|
|
|
|
lines.push(Line::from(vec![
|
2025-08-14 14:10:05 -04:00
|
|
|
|
" • Reasoning Summaries: ".into(),
|
|
|
|
|
|
title_case(&rsum).into(),
|
2025-08-05 23:57:52 -07:00
|
|
|
|
]));
|
|
|
|
|
|
}
|
|
|
|
|
|
|
2025-08-14 14:10:05 -04:00
|
|
|
|
lines.push(Line::from(""));
|
2025-08-07 03:55:59 -07:00
|
|
|
|
|
2025-08-14 14:10:05 -04:00
|
|
|
|
// 📊 Token Usage
|
|
|
|
|
|
lines.push(Line::from(vec!["📊 ".into(), "Token Usage".bold()]));
|
|
|
|
|
|
// Input: <input> [+ <cached> cached]
|
|
|
|
|
|
let mut input_line_spans: Vec<Span<'static>> = vec![
|
|
|
|
|
|
" • Input: ".into(),
|
|
|
|
|
|
usage.non_cached_input().to_string().into(),
|
|
|
|
|
|
];
|
|
|
|
|
|
if let Some(cached) = usage.cached_input_tokens {
|
|
|
|
|
|
if cached > 0 {
|
|
|
|
|
|
input_line_spans.push(format!(" (+ {cached} cached)").into());
|
2025-05-28 14:03:19 -07:00
|
|
|
|
}
|
2025-05-08 21:46:06 -07:00
|
|
|
|
}
|
2025-08-14 14:10:05 -04:00
|
|
|
|
lines.push(Line::from(input_line_spans));
|
|
|
|
|
|
// Output: <output>
|
|
|
|
|
|
lines.push(Line::from(vec![
|
|
|
|
|
|
" • Output: ".into(),
|
|
|
|
|
|
usage.output_tokens.to_string().into(),
|
|
|
|
|
|
]));
|
|
|
|
|
|
// Total: <total>
|
|
|
|
|
|
lines.push(Line::from(vec![
|
|
|
|
|
|
" • Total: ".into(),
|
|
|
|
|
|
usage.blended_total().to_string().into(),
|
|
|
|
|
|
]));
|
|
|
|
|
|
|
|
|
|
|
|
lines.push(Line::from(""));
|
|
|
|
|
|
PlainHistoryCell { lines }
|
|
|
|
|
|
}
|
feat: initial import of Rust implementation of Codex CLI in codex-rs/ (#629)
As stated in `codex-rs/README.md`:
Today, Codex CLI is written in TypeScript and requires Node.js 22+ to
run it. For a number of users, this runtime requirement inhibits
adoption: they would be better served by a standalone executable. As
maintainers, we want Codex to run efficiently in a wide range of
environments with minimal overhead. We also want to take advantage of
operating system-specific APIs to provide better sandboxing, where
possible.
To that end, we are moving forward with a Rust implementation of Codex
CLI contained in this folder, which has the following benefits:
- The CLI compiles to small, standalone, platform-specific binaries.
- Can make direct, native calls to
[seccomp](https://man7.org/linux/man-pages/man2/seccomp.2.html) and
[landlock](https://man7.org/linux/man-pages/man7/landlock.7.html) in
order to support sandboxing on Linux.
- No runtime garbage collection, resulting in lower memory consumption
and better, more predictable performance.
Currently, the Rust implementation is materially behind the TypeScript
implementation in functionality, so continue to use the TypeScript
implmentation for the time being. We will publish native executables via
GitHub Releases as soon as we feel the Rust version is usable.
2025-04-24 13:31:40 -07:00
|
|
|
|
|
2025-08-14 14:10:05 -04:00
|
|
|
|
pub(crate) fn new_prompts_output() -> PlainHistoryCell {
|
|
|
|
|
|
let lines: Vec<Line<'static>> = vec![
|
|
|
|
|
|
Line::from("/prompts".magenta()),
|
|
|
|
|
|
Line::from(""),
|
|
|
|
|
|
Line::from(" 1. Explain this codebase"),
|
|
|
|
|
|
Line::from(" 2. Summarize recent commits"),
|
|
|
|
|
|
Line::from(" 3. Implement {feature}"),
|
|
|
|
|
|
Line::from(" 4. Find and fix a bug in @filename"),
|
|
|
|
|
|
Line::from(" 5. Write tests for @filename"),
|
|
|
|
|
|
Line::from(" 6. Improve documentation in @filename"),
|
|
|
|
|
|
Line::from(""),
|
|
|
|
|
|
];
|
|
|
|
|
|
PlainHistoryCell { lines }
|
|
|
|
|
|
}
|
2025-08-07 00:01:38 -07:00
|
|
|
|
|
2025-08-14 14:10:05 -04:00
|
|
|
|
pub(crate) fn new_error_event(message: String) -> PlainHistoryCell {
|
|
|
|
|
|
let lines: Vec<Line<'static>> = vec![vec!["🖐 ".red().bold(), message.into()].into(), "".into()];
|
|
|
|
|
|
PlainHistoryCell { lines }
|
|
|
|
|
|
}
|
2025-08-07 00:01:38 -07:00
|
|
|
|
|
2025-08-14 14:10:05 -04:00
|
|
|
|
/// Render a user‑friendly plan update styled like a checkbox todo list.
|
|
|
|
|
|
pub(crate) fn new_plan_update(update: UpdatePlanArgs) -> PlainHistoryCell {
|
|
|
|
|
|
let UpdatePlanArgs { explanation, plan } = update;
|
|
|
|
|
|
|
|
|
|
|
|
let mut lines: Vec<Line<'static>> = Vec::new();
|
|
|
|
|
|
// Header with progress summary
|
|
|
|
|
|
let total = plan.len();
|
|
|
|
|
|
let completed = plan
|
|
|
|
|
|
.iter()
|
|
|
|
|
|
.filter(|p| matches!(p.status, StepStatus::Completed))
|
|
|
|
|
|
.count();
|
|
|
|
|
|
|
|
|
|
|
|
let width: usize = 10;
|
|
|
|
|
|
let filled = if total > 0 {
|
|
|
|
|
|
(completed * width + total / 2) / total
|
|
|
|
|
|
} else {
|
|
|
|
|
|
0
|
|
|
|
|
|
};
|
|
|
|
|
|
let empty = width.saturating_sub(filled);
|
|
|
|
|
|
|
|
|
|
|
|
let mut header: Vec<Span> = Vec::new();
|
|
|
|
|
|
header.push(Span::raw("📋"));
|
|
|
|
|
|
header.push(Span::styled(
|
|
|
|
|
|
" Update plan",
|
|
|
|
|
|
Style::default().add_modifier(Modifier::BOLD).magenta(),
|
|
|
|
|
|
));
|
|
|
|
|
|
header.push(Span::raw(" ["));
|
|
|
|
|
|
if filled > 0 {
|
2025-08-07 00:01:38 -07:00
|
|
|
|
header.push(Span::styled(
|
2025-08-14 14:10:05 -04:00
|
|
|
|
"█".repeat(filled),
|
|
|
|
|
|
Style::default().fg(Color::Green),
|
2025-08-07 00:01:38 -07:00
|
|
|
|
));
|
2025-08-14 14:10:05 -04:00
|
|
|
|
}
|
|
|
|
|
|
if empty > 0 {
|
|
|
|
|
|
header.push(Span::styled(
|
|
|
|
|
|
"░".repeat(empty),
|
|
|
|
|
|
Style::default().add_modifier(Modifier::DIM),
|
|
|
|
|
|
));
|
|
|
|
|
|
}
|
|
|
|
|
|
header.push(Span::raw("] "));
|
|
|
|
|
|
header.push(Span::raw(format!("{completed}/{total}")));
|
|
|
|
|
|
lines.push(Line::from(header));
|
|
|
|
|
|
|
|
|
|
|
|
// Optional explanation/note from the model
|
|
|
|
|
|
if let Some(expl) = explanation.and_then(|s| {
|
|
|
|
|
|
let t = s.trim().to_string();
|
|
|
|
|
|
if t.is_empty() { None } else { Some(t) }
|
|
|
|
|
|
}) {
|
|
|
|
|
|
lines.push(Line::from("note".dim().italic()));
|
|
|
|
|
|
for l in expl.lines() {
|
|
|
|
|
|
lines.push(Line::from(l.to_string()).dim());
|
2025-07-31 13:45:52 -07:00
|
|
|
|
}
|
2025-08-14 14:10:05 -04:00
|
|
|
|
}
|
2025-07-31 13:45:52 -07:00
|
|
|
|
|
2025-08-14 14:10:05 -04:00
|
|
|
|
// Steps styled as checkbox items
|
|
|
|
|
|
if plan.is_empty() {
|
|
|
|
|
|
lines.push(Line::from("(no steps provided)".dim().italic()));
|
|
|
|
|
|
} else {
|
|
|
|
|
|
for (idx, PlanItemArg { step, status }) in plan.into_iter().enumerate() {
|
|
|
|
|
|
let (box_span, text_span) = match status {
|
|
|
|
|
|
StepStatus::Completed => (
|
|
|
|
|
|
Span::styled("✔", Style::default().fg(Color::Green)),
|
|
|
|
|
|
Span::styled(
|
|
|
|
|
|
step,
|
|
|
|
|
|
Style::default().add_modifier(Modifier::CROSSED_OUT | Modifier::DIM),
|
2025-08-07 00:01:38 -07:00
|
|
|
|
),
|
2025-08-14 14:10:05 -04:00
|
|
|
|
),
|
|
|
|
|
|
StepStatus::InProgress => (
|
|
|
|
|
|
Span::raw("□"),
|
|
|
|
|
|
Span::styled(
|
|
|
|
|
|
step,
|
|
|
|
|
|
Style::default()
|
|
|
|
|
|
.fg(Color::Blue)
|
|
|
|
|
|
.add_modifier(Modifier::BOLD),
|
2025-08-07 00:01:38 -07:00
|
|
|
|
),
|
2025-08-14 14:10:05 -04:00
|
|
|
|
),
|
|
|
|
|
|
StepStatus::Pending => (
|
|
|
|
|
|
Span::raw("□"),
|
|
|
|
|
|
Span::styled(step, Style::default().add_modifier(Modifier::DIM)),
|
|
|
|
|
|
),
|
|
|
|
|
|
};
|
|
|
|
|
|
let prefix = if idx == 0 {
|
|
|
|
|
|
Span::raw(" └ ")
|
|
|
|
|
|
} else {
|
|
|
|
|
|
Span::raw(" ")
|
|
|
|
|
|
};
|
|
|
|
|
|
lines.push(Line::from(vec![
|
|
|
|
|
|
prefix,
|
|
|
|
|
|
box_span,
|
|
|
|
|
|
Span::raw(" "),
|
|
|
|
|
|
text_span,
|
|
|
|
|
|
]));
|
2025-07-31 13:45:52 -07:00
|
|
|
|
}
|
|
|
|
|
|
}
|
|
|
|
|
|
|
2025-08-14 14:10:05 -04:00
|
|
|
|
lines.push(Line::from(""));
|
2025-08-11 18:32:59 -07:00
|
|
|
|
|
2025-08-14 14:10:05 -04:00
|
|
|
|
PlainHistoryCell { lines }
|
|
|
|
|
|
}
|
feat: initial import of Rust implementation of Codex CLI in codex-rs/ (#629)
As stated in `codex-rs/README.md`:
Today, Codex CLI is written in TypeScript and requires Node.js 22+ to
run it. For a number of users, this runtime requirement inhibits
adoption: they would be better served by a standalone executable. As
maintainers, we want Codex to run efficiently in a wide range of
environments with minimal overhead. We also want to take advantage of
operating system-specific APIs to provide better sandboxing, where
possible.
To that end, we are moving forward with a Rust implementation of Codex
CLI contained in this folder, which has the following benefits:
- The CLI compiles to small, standalone, platform-specific binaries.
- Can make direct, native calls to
[seccomp](https://man7.org/linux/man-pages/man2/seccomp.2.html) and
[landlock](https://man7.org/linux/man-pages/man7/landlock.7.html) in
order to support sandboxing on Linux.
- No runtime garbage collection, resulting in lower memory consumption
and better, more predictable performance.
Currently, the Rust implementation is materially behind the TypeScript
implementation in functionality, so continue to use the TypeScript
implmentation for the time being. We will publish native executables via
GitHub Releases as soon as we feel the Rust version is usable.
2025-04-24 13:31:40 -07:00
|
|
|
|
|
2025-08-14 14:10:05 -04:00
|
|
|
|
/// Create a new `PendingPatch` cell that lists the file‑level summary of
|
|
|
|
|
|
/// a proposed patch. The summary lines should already be formatted (e.g.
|
|
|
|
|
|
/// "A path/to/file.rs").
|
|
|
|
|
|
pub(crate) fn new_patch_event(
|
|
|
|
|
|
event_type: PatchEventType,
|
|
|
|
|
|
changes: HashMap<PathBuf, FileChange>,
|
|
|
|
|
|
) -> PlainHistoryCell {
|
|
|
|
|
|
let title = match &event_type {
|
|
|
|
|
|
PatchEventType::ApprovalRequest => "proposed patch",
|
|
|
|
|
|
PatchEventType::ApplyBegin {
|
|
|
|
|
|
auto_approved: true,
|
|
|
|
|
|
} => "✏️ Applying patch",
|
|
|
|
|
|
PatchEventType::ApplyBegin {
|
|
|
|
|
|
auto_approved: false,
|
|
|
|
|
|
} => {
|
|
|
|
|
|
let lines: Vec<Line<'static>> = vec![
|
|
|
|
|
|
Line::from("✏️ Applying patch".magenta().bold()),
|
|
|
|
|
|
Line::from(""),
|
|
|
|
|
|
];
|
|
|
|
|
|
return PlainHistoryCell { lines };
|
2025-05-28 14:03:19 -07:00
|
|
|
|
}
|
2025-08-14 14:10:05 -04:00
|
|
|
|
};
|
2025-08-05 22:44:27 -07:00
|
|
|
|
|
2025-08-14 14:10:05 -04:00
|
|
|
|
let mut lines: Vec<Line<'static>> = create_diff_summary(title, &changes, event_type);
|
2025-08-06 12:03:45 -07:00
|
|
|
|
|
2025-08-14 14:10:05 -04:00
|
|
|
|
lines.push(Line::from(""));
|
2025-08-05 22:44:27 -07:00
|
|
|
|
|
2025-08-14 14:10:05 -04:00
|
|
|
|
PlainHistoryCell { lines }
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
pub(crate) fn new_patch_apply_failure(stderr: String) -> PlainHistoryCell {
|
|
|
|
|
|
let mut lines: Vec<Line<'static>> = Vec::new();
|
|
|
|
|
|
|
|
|
|
|
|
// Failure title
|
|
|
|
|
|
lines.push(Line::from("✘ Failed to apply patch".magenta().bold()));
|
|
|
|
|
|
|
|
|
|
|
|
if !stderr.trim().is_empty() {
|
|
|
|
|
|
lines.extend(output_lines(
|
|
|
|
|
|
Some(&CommandOutput {
|
|
|
|
|
|
exit_code: 1,
|
|
|
|
|
|
stdout: String::new(),
|
|
|
|
|
|
stderr,
|
|
|
|
|
|
}),
|
|
|
|
|
|
true,
|
|
|
|
|
|
true,
|
|
|
|
|
|
));
|
2025-08-05 22:44:27 -07:00
|
|
|
|
}
|
2025-08-12 17:37:28 -07:00
|
|
|
|
|
2025-08-14 14:10:05 -04:00
|
|
|
|
lines.push(Line::from(""));
|
2025-08-12 17:37:28 -07:00
|
|
|
|
|
2025-08-14 14:10:05 -04:00
|
|
|
|
PlainHistoryCell { lines }
|
|
|
|
|
|
}
|
2025-08-12 17:37:28 -07:00
|
|
|
|
|
2025-08-14 14:10:05 -04:00
|
|
|
|
pub(crate) fn new_patch_apply_success(stdout: String) -> PlainHistoryCell {
|
|
|
|
|
|
let mut lines: Vec<Line<'static>> = Vec::new();
|
2025-08-12 17:37:28 -07:00
|
|
|
|
|
2025-08-14 14:10:05 -04:00
|
|
|
|
// Success title
|
|
|
|
|
|
lines.push(Line::from("✓ Applied patch".magenta().bold()));
|
2025-08-12 17:37:28 -07:00
|
|
|
|
|
2025-08-14 14:10:05 -04:00
|
|
|
|
if !stdout.trim().is_empty() {
|
|
|
|
|
|
let mut iter = stdout.lines();
|
|
|
|
|
|
for (i, raw) in iter.by_ref().take(TOOL_CALL_MAX_LINES).enumerate() {
|
|
|
|
|
|
let prefix = if i == 0 { " └ " } else { " " };
|
|
|
|
|
|
let s = format!("{prefix}{raw}");
|
|
|
|
|
|
lines.push(ansi_escape_line(&s).dim());
|
|
|
|
|
|
}
|
|
|
|
|
|
let remaining = iter.count();
|
|
|
|
|
|
if remaining > 0 {
|
|
|
|
|
|
lines.push(Line::from(""));
|
|
|
|
|
|
lines.push(Line::from(format!("... +{remaining} lines")).dim());
|
2025-08-12 17:37:28 -07:00
|
|
|
|
}
|
|
|
|
|
|
}
|
2025-05-28 14:03:19 -07:00
|
|
|
|
|
2025-08-14 14:10:05 -04:00
|
|
|
|
lines.push(Line::from(""));
|
|
|
|
|
|
|
|
|
|
|
|
PlainHistoryCell { lines }
|
2025-08-06 12:03:45 -07:00
|
|
|
|
}
|
|
|
|
|
|
|
2025-08-11 11:26:15 -07:00
|
|
|
|
fn output_lines(
|
|
|
|
|
|
output: Option<&CommandOutput>,
|
|
|
|
|
|
only_err: bool,
|
|
|
|
|
|
include_angle_pipe: bool,
|
|
|
|
|
|
) -> Vec<Line<'static>> {
|
|
|
|
|
|
let CommandOutput {
|
|
|
|
|
|
exit_code,
|
|
|
|
|
|
stdout,
|
|
|
|
|
|
stderr,
|
|
|
|
|
|
} = match output {
|
|
|
|
|
|
Some(output) if only_err && output.exit_code == 0 => return vec![],
|
|
|
|
|
|
Some(output) => output,
|
|
|
|
|
|
None => return vec![],
|
|
|
|
|
|
};
|
|
|
|
|
|
|
|
|
|
|
|
let src = if *exit_code == 0 { stdout } else { stderr };
|
|
|
|
|
|
let lines: Vec<&str> = src.lines().collect();
|
|
|
|
|
|
let total = lines.len();
|
|
|
|
|
|
let limit = TOOL_CALL_MAX_LINES;
|
|
|
|
|
|
|
|
|
|
|
|
let mut out = Vec::new();
|
|
|
|
|
|
|
|
|
|
|
|
let head_end = total.min(limit);
|
|
|
|
|
|
for (i, raw) in lines[..head_end].iter().enumerate() {
|
|
|
|
|
|
let mut line = ansi_escape_line(raw);
|
|
|
|
|
|
let prefix = if i == 0 && include_angle_pipe {
|
2025-08-13 19:14:03 -04:00
|
|
|
|
" └ "
|
2025-08-11 11:26:15 -07:00
|
|
|
|
} else {
|
|
|
|
|
|
" "
|
|
|
|
|
|
};
|
|
|
|
|
|
line.spans.insert(0, prefix.into());
|
|
|
|
|
|
line.spans.iter_mut().for_each(|span| {
|
|
|
|
|
|
span.style = span.style.add_modifier(Modifier::DIM);
|
|
|
|
|
|
});
|
|
|
|
|
|
out.push(line);
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
// If we will ellipsize less than the limit, just show it.
|
|
|
|
|
|
let show_ellipsis = total > 2 * limit;
|
|
|
|
|
|
if show_ellipsis {
|
|
|
|
|
|
let omitted = total - 2 * limit;
|
|
|
|
|
|
out.push(Line::from(format!("… +{omitted} lines")));
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
let tail_start = if show_ellipsis {
|
|
|
|
|
|
total - limit
|
|
|
|
|
|
} else {
|
|
|
|
|
|
head_end
|
|
|
|
|
|
};
|
|
|
|
|
|
for raw in lines[tail_start..].iter() {
|
|
|
|
|
|
let mut line = ansi_escape_line(raw);
|
|
|
|
|
|
line.spans.insert(0, " ".into());
|
|
|
|
|
|
line.spans.iter_mut().for_each(|span| {
|
2025-08-06 22:25:41 -07:00
|
|
|
|
span.style = span.style.add_modifier(Modifier::DIM);
|
|
|
|
|
|
});
|
|
|
|
|
|
out.push(line);
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
out
|
feat: initial import of Rust implementation of Codex CLI in codex-rs/ (#629)
As stated in `codex-rs/README.md`:
Today, Codex CLI is written in TypeScript and requires Node.js 22+ to
run it. For a number of users, this runtime requirement inhibits
adoption: they would be better served by a standalone executable. As
maintainers, we want Codex to run efficiently in a wide range of
environments with minimal overhead. We also want to take advantage of
operating system-specific APIs to provide better sandboxing, where
possible.
To that end, we are moving forward with a Rust implementation of Codex
CLI contained in this folder, which has the following benefits:
- The CLI compiles to small, standalone, platform-specific binaries.
- Can make direct, native calls to
[seccomp](https://man7.org/linux/man-pages/man2/seccomp.2.html) and
[landlock](https://man7.org/linux/man-pages/man7/landlock.7.html) in
order to support sandboxing on Linux.
- No runtime garbage collection, resulting in lower memory consumption
and better, more predictable performance.
Currently, the Rust implementation is materially behind the TypeScript
implementation in functionality, so continue to use the TypeScript
implmentation for the time being. We will publish native executables via
GitHub Releases as soon as we feel the Rust version is usable.
2025-04-24 13:31:40 -07:00
|
|
|
|
}
|
fix: introduce ResponseInputItem::McpToolCallOutput variant (#1151)
The output of an MCP server tool call can be one of several types, but
to date, we treated all outputs as text by showing the serialized JSON
as the "tool output" in Codex:
https://github.com/openai/codex/blob/25a9949c49194d5a64de54a11bcc5b4724ac9bd5/codex-rs/mcp-types/src/lib.rs#L96-L101
This PR adds support for the `ImageContent` variant so we can now
display an image output from an MCP tool call.
In making this change, we introduce a new
`ResponseInputItem::McpToolCallOutput` variant so that we can work with
the `mcp_types::CallToolResult` directly when the function call is made
to an MCP server.
Though arguably the more significant change is the introduction of
`HistoryCell::CompletedMcpToolCallWithImageOutput`, which is a cell that
uses `ratatui_image` to render an image into the terminal. To support
this, we introduce `ImageRenderCache`, cache a
`ratatui_image::picker::Picker`, and `ensure_image_cache()` to cache the
appropriate scaled image data and dimensions based on the current
terminal size.
To test, I created a minimal `package.json`:
```json
{
"name": "kitty-mcp",
"version": "1.0.0",
"type": "module",
"description": "MCP that returns image of kitty",
"main": "index.js",
"dependencies": {
"@modelcontextprotocol/sdk": "^1.12.0"
}
}
```
with the following `index.js` to define the MCP server:
```js
#!/usr/bin/env node
import { McpServer } from "@modelcontextprotocol/sdk/server/mcp.js";
import { StdioServerTransport } from "@modelcontextprotocol/sdk/server/stdio.js";
import { readFile } from "node:fs/promises";
import { join } from "node:path";
const IMAGE_URI = "image://Ada.png";
const server = new McpServer({
name: "Demo",
version: "1.0.0",
});
server.tool(
"get-cat-image",
"If you need a cat image, this tool will provide one.",
async () => ({
content: [
{ type: "image", data: await getAdaPngBase64(), mimeType: "image/png" },
],
})
);
server.resource("Ada the Cat", IMAGE_URI, async (uri) => {
const base64Image = await getAdaPngBase64();
return {
contents: [
{
uri: uri.href,
mimeType: "image/png",
blob: base64Image,
},
],
};
});
async function getAdaPngBase64() {
const __dirname = new URL(".", import.meta.url).pathname;
// From https://github.com/benjajaja/ratatui-image/blob/9705ce2c59ec669abbce2924cbfd1f5ae22c9860/assets/Ada.png
const filePath = join(__dirname, "Ada.png");
const imageData = await readFile(filePath);
const base64Image = imageData.toString("base64");
return base64Image;
}
const transport = new StdioServerTransport();
await server.connect(transport);
```
With the local changes from this PR, I added the following to my
`config.toml`:
```toml
[mcp_servers.kitty]
command = "node"
args = ["/Users/mbolin/code/kitty-mcp/index.js"]
```
Running the TUI from source:
```
cargo run --bin codex -- --model o3 'I need a picture of a cat'
```
I get:
<img width="732" alt="image"
src="https://github.com/user-attachments/assets/bf80b721-9ca0-4d81-aec7-77d6899e2869"
/>
Now, that said, I have only tested in iTerm and there is definitely some
funny business with getting an accurate character-to-pixel ratio
(sometimes the `CompletedMcpToolCallWithImageOutput` thinks it needs 10
rows to render instead of 4), so there is still work to be done here.
2025-05-28 19:03:17 -07:00
|
|
|
|
|
2025-07-30 10:05:40 -07:00
|
|
|
|
fn format_mcp_invocation<'a>(invocation: McpInvocation) -> Line<'a> {
|
|
|
|
|
|
let args_str = invocation
|
|
|
|
|
|
.arguments
|
|
|
|
|
|
.as_ref()
|
|
|
|
|
|
.map(|v| {
|
|
|
|
|
|
// Use compact form to keep things short but readable.
|
|
|
|
|
|
serde_json::to_string(v).unwrap_or_else(|_| v.to_string())
|
|
|
|
|
|
})
|
|
|
|
|
|
.unwrap_or_default();
|
|
|
|
|
|
|
|
|
|
|
|
let invocation_spans = vec![
|
|
|
|
|
|
Span::styled(invocation.server.clone(), Style::default().fg(Color::Blue)),
|
|
|
|
|
|
Span::raw("."),
|
|
|
|
|
|
Span::styled(invocation.tool.clone(), Style::default().fg(Color::Blue)),
|
|
|
|
|
|
Span::raw("("),
|
2025-08-13 15:50:50 -07:00
|
|
|
|
Span::styled(args_str, Style::default().add_modifier(Modifier::DIM)),
|
2025-07-30 10:05:40 -07:00
|
|
|
|
Span::raw(")"),
|
|
|
|
|
|
];
|
|
|
|
|
|
Line::from(invocation_spans)
|
fix: introduce ResponseInputItem::McpToolCallOutput variant (#1151)
The output of an MCP server tool call can be one of several types, but
to date, we treated all outputs as text by showing the serialized JSON
as the "tool output" in Codex:
https://github.com/openai/codex/blob/25a9949c49194d5a64de54a11bcc5b4724ac9bd5/codex-rs/mcp-types/src/lib.rs#L96-L101
This PR adds support for the `ImageContent` variant so we can now
display an image output from an MCP tool call.
In making this change, we introduce a new
`ResponseInputItem::McpToolCallOutput` variant so that we can work with
the `mcp_types::CallToolResult` directly when the function call is made
to an MCP server.
Though arguably the more significant change is the introduction of
`HistoryCell::CompletedMcpToolCallWithImageOutput`, which is a cell that
uses `ratatui_image` to render an image into the terminal. To support
this, we introduce `ImageRenderCache`, cache a
`ratatui_image::picker::Picker`, and `ensure_image_cache()` to cache the
appropriate scaled image data and dimensions based on the current
terminal size.
To test, I created a minimal `package.json`:
```json
{
"name": "kitty-mcp",
"version": "1.0.0",
"type": "module",
"description": "MCP that returns image of kitty",
"main": "index.js",
"dependencies": {
"@modelcontextprotocol/sdk": "^1.12.0"
}
}
```
with the following `index.js` to define the MCP server:
```js
#!/usr/bin/env node
import { McpServer } from "@modelcontextprotocol/sdk/server/mcp.js";
import { StdioServerTransport } from "@modelcontextprotocol/sdk/server/stdio.js";
import { readFile } from "node:fs/promises";
import { join } from "node:path";
const IMAGE_URI = "image://Ada.png";
const server = new McpServer({
name: "Demo",
version: "1.0.0",
});
server.tool(
"get-cat-image",
"If you need a cat image, this tool will provide one.",
async () => ({
content: [
{ type: "image", data: await getAdaPngBase64(), mimeType: "image/png" },
],
})
);
server.resource("Ada the Cat", IMAGE_URI, async (uri) => {
const base64Image = await getAdaPngBase64();
return {
contents: [
{
uri: uri.href,
mimeType: "image/png",
blob: base64Image,
},
],
};
});
async function getAdaPngBase64() {
const __dirname = new URL(".", import.meta.url).pathname;
// From https://github.com/benjajaja/ratatui-image/blob/9705ce2c59ec669abbce2924cbfd1f5ae22c9860/assets/Ada.png
const filePath = join(__dirname, "Ada.png");
const imageData = await readFile(filePath);
const base64Image = imageData.toString("base64");
return base64Image;
}
const transport = new StdioServerTransport();
await server.connect(transport);
```
With the local changes from this PR, I added the following to my
`config.toml`:
```toml
[mcp_servers.kitty]
command = "node"
args = ["/Users/mbolin/code/kitty-mcp/index.js"]
```
Running the TUI from source:
```
cargo run --bin codex -- --model o3 'I need a picture of a cat'
```
I get:
<img width="732" alt="image"
src="https://github.com/user-attachments/assets/bf80b721-9ca0-4d81-aec7-77d6899e2869"
/>
Now, that said, I have only tested in iTerm and there is definitely some
funny business with getting an accurate character-to-pixel ratio
(sometimes the `CompletedMcpToolCallWithImageOutput` thinks it needs 10
rows to render instead of 4), so there is still work to be done here.
2025-05-28 19:03:17 -07:00
|
|
|
|
}
|
2025-08-11 11:26:15 -07:00
|
|
|
|
|
|
|
|
|
|
fn shlex_join_safe(command: &[String]) -> String {
|
|
|
|
|
|
match shlex::try_join(command.iter().map(|s| s.as_str())) {
|
|
|
|
|
|
Ok(cmd) => cmd,
|
|
|
|
|
|
Err(_) => command.join(" "),
|
|
|
|
|
|
}
|
|
|
|
|
|
}
|
2025-08-11 16:11:46 -07:00
|
|
|
|
|
|
|
|
|
|
#[cfg(test)]
|
|
|
|
|
|
mod tests {
|
|
|
|
|
|
use super::*;
|
|
|
|
|
|
|
|
|
|
|
|
#[test]
|
|
|
|
|
|
fn parsed_command_with_newlines_starts_each_line_at_origin() {
|
|
|
|
|
|
let parsed = vec![ParsedCommand::Unknown {
|
|
|
|
|
|
cmd: vec!["printf".into(), "foo\nbar".into()],
|
|
|
|
|
|
}];
|
2025-08-14 14:10:05 -04:00
|
|
|
|
let lines = exec_command_lines(&[], &parsed, None);
|
2025-08-11 16:11:46 -07:00
|
|
|
|
assert!(lines.len() >= 3);
|
2025-08-13 19:14:03 -04:00
|
|
|
|
assert_eq!(lines[1].spans[0].content, " └ ");
|
2025-08-11 16:11:46 -07:00
|
|
|
|
assert_eq!(lines[2].spans[0].content, " ");
|
|
|
|
|
|
}
|
|
|
|
|
|
}
|