mirror of
https://github.com/rojo-rbx/rojo.git
synced 2026-04-23 06:05:24 +00:00
Compare commits
15 Commits
| Author | SHA1 | Date | |
|---|---|---|---|
| a6e9939d6c | |||
|
5957368c04
|
|||
|
78916c8a63
|
|||
|
791ccfcfd1
|
|||
|
3500ebe02a
|
|||
|
0e1364945f
|
|||
| 3a6aae65f7 | |||
| d13d229eef | |||
| 9a485d88ce | |||
|
020d72faef
|
|||
|
60d150f4c6
|
|||
|
73dab330b5
|
|||
|
790312a5b0
|
|||
|
5c396322d9
|
|||
|
37e44e474a
|
7
.github/workflows/release.yml
vendored
7
.github/workflows/release.yml
vendored
@@ -45,13 +45,6 @@ jobs:
|
|||||||
name: Rojo.rbxm
|
name: Rojo.rbxm
|
||||||
path: Rojo.rbxm
|
path: Rojo.rbxm
|
||||||
|
|
||||||
- name: Upload Plugin to Roblox
|
|
||||||
env:
|
|
||||||
RBX_API_KEY: ${{ secrets.PLUGIN_UPLOAD_TOKEN }}
|
|
||||||
RBX_UNIVERSE_ID: ${{ vars.PLUGIN_CI_PLACE_ID }}
|
|
||||||
RBX_PLACE_ID: ${{ vars.PLUGIN_CI_UNIVERSE_ID }}
|
|
||||||
run: lune run upload-plugin Rojo.rbxm
|
|
||||||
|
|
||||||
build:
|
build:
|
||||||
needs: ["create-release"]
|
needs: ["create-release"]
|
||||||
strategy:
|
strategy:
|
||||||
|
|||||||
6
.gitmodules
vendored
6
.gitmodules
vendored
@@ -16,9 +16,3 @@
|
|||||||
[submodule "plugin/Packages/Highlighter"]
|
[submodule "plugin/Packages/Highlighter"]
|
||||||
path = plugin/Packages/Highlighter
|
path = plugin/Packages/Highlighter
|
||||||
url = https://github.com/boatbomber/highlighter.git
|
url = https://github.com/boatbomber/highlighter.git
|
||||||
[submodule "plugin/Packages/msgpack-luau"]
|
|
||||||
path = plugin/Packages/msgpack-luau
|
|
||||||
url = https://github.com/cipharius/msgpack-luau/
|
|
||||||
[submodule ".lune/opencloud-execute"]
|
|
||||||
path = .lune/opencloud-execute
|
|
||||||
url = https://github.com/Dekkonot/opencloud-luau-execute-lune.git
|
|
||||||
|
|||||||
@@ -1,8 +0,0 @@
|
|||||||
return {
|
|
||||||
luau = {
|
|
||||||
languagemode = "strict",
|
|
||||||
aliases = {
|
|
||||||
lune = "~/.lune/.typedefs/0.10.4/",
|
|
||||||
},
|
|
||||||
},
|
|
||||||
}
|
|
||||||
Submodule .lune/opencloud-execute deleted from 8ae86dd3ad
@@ -1,51 +0,0 @@
|
|||||||
local args: any = ...
|
|
||||||
assert(args, "no arguments passed to script")
|
|
||||||
|
|
||||||
local input: buffer = args.BinaryInput
|
|
||||||
|
|
||||||
local AssetService = game:GetService("AssetService")
|
|
||||||
local SerializationService = game:GetService("SerializationService")
|
|
||||||
local EncodingService = game:GetService("EncodingService")
|
|
||||||
|
|
||||||
local input_hash: buffer = EncodingService:ComputeBufferHash(input, Enum.HashAlgorithm.Sha256)
|
|
||||||
local hex_hash: { string } = table.create(buffer.len(input_hash))
|
|
||||||
for i = 0, buffer.len(input_hash) - 1 do
|
|
||||||
table.insert(hex_hash, string.format("%02x", buffer.readu8(input_hash, i)))
|
|
||||||
end
|
|
||||||
|
|
||||||
print(`Deserializing plugin file (size: {buffer.len(input)} bytes, hash: {table.concat(hex_hash, "")})`)
|
|
||||||
local plugin = SerializationService:DeserializeInstancesAsync(input)[1]
|
|
||||||
|
|
||||||
local UploadDetails = require(plugin.UploadDetails) :: any
|
|
||||||
local PLUGIN_ID = UploadDetails.assetId
|
|
||||||
local PLUGIN_NAME = UploadDetails.name
|
|
||||||
local PLUGIN_DESCRIPTION = UploadDetails.description
|
|
||||||
local PLUGIN_CREATOR_ID = UploadDetails.creatorId
|
|
||||||
local PLUGIN_CREATOR_TYPE = UploadDetails.creatorType
|
|
||||||
|
|
||||||
assert(typeof(PLUGIN_ID) == "number", "UploadDetails did not contain a number field 'assetId'")
|
|
||||||
assert(typeof(PLUGIN_NAME) == "string", "UploadDetails did not contain a string field 'name'")
|
|
||||||
assert(typeof(PLUGIN_DESCRIPTION) == "string", "UploadDetails did not contain a string field 'description'")
|
|
||||||
assert(typeof(PLUGIN_CREATOR_ID) == "number", "UploadDetails did not contain a number field 'creatorId'")
|
|
||||||
assert(typeof(PLUGIN_CREATOR_TYPE) == "string", "UploadDetails did not contain a string field 'creatorType'")
|
|
||||||
assert(
|
|
||||||
Enum.AssetCreatorType:FromName(PLUGIN_CREATOR_TYPE) ~= nil,
|
|
||||||
"UploadDetails field 'creatorType' was not a valid member of Enum.AssetCreatorType"
|
|
||||||
)
|
|
||||||
|
|
||||||
print(`Uploading to {PLUGIN_ID}`)
|
|
||||||
print(`Plugin Name: {PLUGIN_NAME}`)
|
|
||||||
print(`Plugin Description: {PLUGIN_DESCRIPTION}`)
|
|
||||||
|
|
||||||
local result, version_or_err = AssetService:CreateAssetVersionAsync(plugin, Enum.AssetType.Plugin, PLUGIN_ID, {
|
|
||||||
["Name"] = PLUGIN_NAME,
|
|
||||||
["Description"] = PLUGIN_DESCRIPTION,
|
|
||||||
["CreatorId"] = PLUGIN_CREATOR_ID,
|
|
||||||
["CreatorType"] = Enum.AssetCreatorType:FromName(PLUGIN_CREATOR_TYPE),
|
|
||||||
})
|
|
||||||
|
|
||||||
if result ~= Enum.CreateAssetResult.Success then
|
|
||||||
error(`Plugin failed to upload because: {result.Name} - {version_or_err}`)
|
|
||||||
end
|
|
||||||
|
|
||||||
print(`Plugin uploaded successfully. New version is {version_or_err}.`)
|
|
||||||
@@ -1,78 +0,0 @@
|
|||||||
local fs = require("@lune/fs")
|
|
||||||
local process = require("@lune/process")
|
|
||||||
local stdio = require("@lune/stdio")
|
|
||||||
|
|
||||||
local luau_execute = require("./opencloud-execute")
|
|
||||||
|
|
||||||
local UNIVERSE_ID = process.env["RBX_UNIVERSE_ID"]
|
|
||||||
local PLACE_ID = process.env["RBX_PLACE_ID"]
|
|
||||||
|
|
||||||
local version_string = fs.readFile("plugin/Version.txt")
|
|
||||||
local versions = { string.match(version_string, "^v?(%d+)%.(%d+)%.(%d+)(.*)$") }
|
|
||||||
if versions[4] ~= "" then
|
|
||||||
print("This release is a pre-release. Skipping uploading plugin.")
|
|
||||||
process.exit(0)
|
|
||||||
end
|
|
||||||
|
|
||||||
local plugin_path = process.args[1]
|
|
||||||
assert(
|
|
||||||
typeof(plugin_path) == "string",
|
|
||||||
"no plugin path provided, expected usage is `lune run upload-plugin [PATH TO RBXM]`."
|
|
||||||
)
|
|
||||||
|
|
||||||
-- For local testing
|
|
||||||
if process.env["CI"] ~= "true" then
|
|
||||||
local rojo = process.exec("rojo", { "build", "plugin.project.json", "--output", plugin_path })
|
|
||||||
if not rojo.ok then
|
|
||||||
stdio.ewrite("plugin upload failed because: could not build plugin.rbxm\n\n")
|
|
||||||
stdio.ewrite(rojo.stderr)
|
|
||||||
stdio.ewrite("\n")
|
|
||||||
process.exit(1)
|
|
||||||
end
|
|
||||||
else
|
|
||||||
assert(fs.isFile(plugin_path), `Plugin file did not exist at {plugin_path}`)
|
|
||||||
end
|
|
||||||
local plugin_content = fs.readFile(plugin_path)
|
|
||||||
|
|
||||||
local engine_script = fs.readFile(".lune/scripts/plugin-upload.luau")
|
|
||||||
|
|
||||||
print("Creating task to upload plugin")
|
|
||||||
local task = luau_execute.create_task_latest(UNIVERSE_ID, PLACE_ID, engine_script, 300, false, plugin_content)
|
|
||||||
|
|
||||||
print("Waiting for task to finish")
|
|
||||||
local success = luau_execute.await_finish(task)
|
|
||||||
if not success then
|
|
||||||
local error = luau_execute.get_error(task)
|
|
||||||
assert(error, "could not fetch error from task")
|
|
||||||
stdio.ewrite("plugin upload failed because: task did not finish successfully\n\n")
|
|
||||||
stdio.ewrite(error.code)
|
|
||||||
stdio.ewrite("\n")
|
|
||||||
stdio.ewrite(error.message)
|
|
||||||
stdio.ewrite("\n")
|
|
||||||
process.exit(1)
|
|
||||||
end
|
|
||||||
|
|
||||||
print("Output from task:\n")
|
|
||||||
for _, log in luau_execute.get_structured_logs(task) do
|
|
||||||
if log.messageType == "ERROR" then
|
|
||||||
stdio.write(stdio.color("red"))
|
|
||||||
stdio.write(log.message)
|
|
||||||
stdio.write("\n")
|
|
||||||
stdio.write(stdio.color("reset"))
|
|
||||||
elseif log.messageType == "INFO" then
|
|
||||||
stdio.write(stdio.color("cyan"))
|
|
||||||
stdio.write(log.message)
|
|
||||||
stdio.write("\n")
|
|
||||||
stdio.write(stdio.color("reset"))
|
|
||||||
elseif log.messageType == "WARNING" then
|
|
||||||
stdio.write(stdio.color("yellow"))
|
|
||||||
stdio.write(log.message)
|
|
||||||
stdio.write("\n")
|
|
||||||
stdio.write(stdio.color("reset"))
|
|
||||||
else
|
|
||||||
stdio.write(stdio.color("reset"))
|
|
||||||
stdio.write(log.message)
|
|
||||||
stdio.write("\n")
|
|
||||||
stdio.write(stdio.color("reset"))
|
|
||||||
end
|
|
||||||
end
|
|
||||||
13
CHANGELOG.md
13
CHANGELOG.md
@@ -30,24 +30,15 @@ Making a new release? Simply add the new header with the version and date undern
|
|||||||
-->
|
-->
|
||||||
|
|
||||||
## Unreleased
|
## Unreleased
|
||||||
|
|
||||||
* `inf` and `nan` values in properties are now synced ([#1176])
|
|
||||||
* Fixed a bug caused by having reference properties (such as `ObjectValue.Value`) that point to an Instance not included in syncback. ([#1179])
|
* Fixed a bug caused by having reference properties (such as `ObjectValue.Value`) that point to an Instance not included in syncback. ([#1179])
|
||||||
|
* Implemented support for the "name" property in meta/model JSON files. ([#1187])
|
||||||
* Fixed instance replacement fallback failing when too many instances needed to be replaced. ([#1192])
|
* Fixed instance replacement fallback failing when too many instances needed to be replaced. ([#1192])
|
||||||
* Added actors and bindable/remote event/function variants to be synced back as JSON files. ([#1199])
|
|
||||||
* Fixed a bug where MacOS paths weren't being handled correctly. ([#1201])
|
* Fixed a bug where MacOS paths weren't being handled correctly. ([#1201])
|
||||||
* Fixed a bug where the notification timeout thread would fail to cancel on unmount ([#1211])
|
|
||||||
* Added a "Forget" option to the sync reminder notification to avoid being reminded for that place in the future ([#1215])
|
|
||||||
* Improves relative path calculation for sourcemap generation to avoid issues with Windows UNC paths. ([#1217])
|
|
||||||
|
|
||||||
[#1176]: https://github.com/rojo-rbx/rojo/pull/1176
|
|
||||||
[#1179]: https://github.com/rojo-rbx/rojo/pull/1179
|
[#1179]: https://github.com/rojo-rbx/rojo/pull/1179
|
||||||
|
[#1187]: https://github.com/rojo-rbx/rojo/pull/1187
|
||||||
[#1192]: https://github.com/rojo-rbx/rojo/pull/1192
|
[#1192]: https://github.com/rojo-rbx/rojo/pull/1192
|
||||||
[#1199]: https://github.com/rojo-rbx/rojo/pull/1199
|
|
||||||
[#1201]: https://github.com/rojo-rbx/rojo/pull/1201
|
[#1201]: https://github.com/rojo-rbx/rojo/pull/1201
|
||||||
[#1211]: https://github.com/rojo-rbx/rojo/pull/1211
|
|
||||||
[#1215]: https://github.com/rojo-rbx/rojo/pull/1215
|
|
||||||
[#1217]: https://github.com/rojo-rbx/rojo/pull/1217
|
|
||||||
|
|
||||||
## [7.7.0-rc.1] (November 27th, 2025)
|
## [7.7.0-rc.1] (November 27th, 2025)
|
||||||
|
|
||||||
|
|||||||
@@ -14,7 +14,6 @@ Code contributions are welcome for features and bugs that have been reported in
|
|||||||
You'll want these tools to work on Rojo:
|
You'll want these tools to work on Rojo:
|
||||||
|
|
||||||
* Latest stable Rust compiler
|
* Latest stable Rust compiler
|
||||||
* Rustfmt and Clippy are used for code formatting and linting.
|
|
||||||
* Latest stable [Rojo](https://github.com/rojo-rbx/rojo)
|
* Latest stable [Rojo](https://github.com/rojo-rbx/rojo)
|
||||||
* [Rokit](https://github.com/rojo-rbx/rokit)
|
* [Rokit](https://github.com/rojo-rbx/rokit)
|
||||||
* [Luau Language Server](https://github.com/JohnnyMorganz/luau-lsp) (Only needed if working on the Studio plugin.)
|
* [Luau Language Server](https://github.com/JohnnyMorganz/luau-lsp) (Only needed if working on the Studio plugin.)
|
||||||
|
|||||||
19
Cargo.lock
generated
19
Cargo.lock
generated
@@ -1520,12 +1520,6 @@ version = "1.0.15"
|
|||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "57c0d7b74b563b49d38dae00a0c37d4d6de9b432382b2892f0574ddcae73fd0a"
|
checksum = "57c0d7b74b563b49d38dae00a0c37d4d6de9b432382b2892f0574ddcae73fd0a"
|
||||||
|
|
||||||
[[package]]
|
|
||||||
name = "pathdiff"
|
|
||||||
version = "0.2.3"
|
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
|
||||||
checksum = "df94ce210e5bc13cb6651479fa48d14f601d9858cfe0467f43ae157023b938d3"
|
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "percent-encoding"
|
name = "percent-encoding"
|
||||||
version = "2.3.2"
|
version = "2.3.2"
|
||||||
@@ -2074,7 +2068,6 @@ dependencies = [
|
|||||||
"num_cpus",
|
"num_cpus",
|
||||||
"opener",
|
"opener",
|
||||||
"paste",
|
"paste",
|
||||||
"pathdiff",
|
|
||||||
"pretty_assertions",
|
"pretty_assertions",
|
||||||
"profiling",
|
"profiling",
|
||||||
"rayon",
|
"rayon",
|
||||||
@@ -2085,12 +2078,10 @@ dependencies = [
|
|||||||
"rbx_xml",
|
"rbx_xml",
|
||||||
"reqwest",
|
"reqwest",
|
||||||
"ritz",
|
"ritz",
|
||||||
"rmp-serde",
|
|
||||||
"roblox_install",
|
"roblox_install",
|
||||||
"rojo-insta-ext",
|
"rojo-insta-ext",
|
||||||
"semver",
|
"semver",
|
||||||
"serde",
|
"serde",
|
||||||
"serde_bytes",
|
|
||||||
"serde_json",
|
"serde_json",
|
||||||
"serde_yaml",
|
"serde_yaml",
|
||||||
"strum",
|
"strum",
|
||||||
@@ -2231,16 +2222,6 @@ dependencies = [
|
|||||||
"serde_derive",
|
"serde_derive",
|
||||||
]
|
]
|
||||||
|
|
||||||
[[package]]
|
|
||||||
name = "serde_bytes"
|
|
||||||
version = "0.11.19"
|
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
|
||||||
checksum = "a5d440709e79d88e51ac01c4b72fc6cb7314017bb7da9eeff678aa94c10e3ea8"
|
|
||||||
dependencies = [
|
|
||||||
"serde",
|
|
||||||
"serde_core",
|
|
||||||
]
|
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "serde_cbor"
|
name = "serde_cbor"
|
||||||
version = "0.11.2"
|
version = "0.11.2"
|
||||||
|
|||||||
@@ -100,13 +100,10 @@ clap = { version = "3.2.25", features = ["derive"] }
|
|||||||
profiling = "1.0.15"
|
profiling = "1.0.15"
|
||||||
yaml-rust2 = "0.10.3"
|
yaml-rust2 = "0.10.3"
|
||||||
data-encoding = "2.8.0"
|
data-encoding = "2.8.0"
|
||||||
pathdiff = "0.2.3"
|
|
||||||
|
|
||||||
blake3 = "1.5.0"
|
blake3 = "1.5.0"
|
||||||
float-cmp = "0.9.0"
|
float-cmp = "0.9.0"
|
||||||
indexmap = { version = "2.10.0", features = ["serde"] }
|
indexmap = { version = "2.10.0", features = ["serde"] }
|
||||||
rmp-serde = "1.3.0"
|
|
||||||
serde_bytes = "0.11.19"
|
|
||||||
|
|
||||||
[target.'cfg(windows)'.dependencies]
|
[target.'cfg(windows)'.dependencies]
|
||||||
winreg = "0.10.1"
|
winreg = "0.10.1"
|
||||||
@@ -125,7 +122,7 @@ semver = "1.0.22"
|
|||||||
rojo-insta-ext = { path = "crates/rojo-insta-ext" }
|
rojo-insta-ext = { path = "crates/rojo-insta-ext" }
|
||||||
|
|
||||||
criterion = "0.3.6"
|
criterion = "0.3.6"
|
||||||
insta = { version = "1.36.1", features = ["redactions", "yaml", "json"] }
|
insta = { version = "1.36.1", features = ["redactions", "yaml"] }
|
||||||
paste = "1.0.14"
|
paste = "1.0.14"
|
||||||
pretty_assertions = "1.4.0"
|
pretty_assertions = "1.4.0"
|
||||||
serde_yaml = "0.8.26"
|
serde_yaml = "0.8.26"
|
||||||
|
|||||||
6
build.rs
6
build.rs
@@ -30,11 +30,6 @@ fn snapshot_from_fs_path(path: &Path) -> io::Result<VfsSnapshot> {
|
|||||||
continue;
|
continue;
|
||||||
}
|
}
|
||||||
|
|
||||||
// Ignore images in msgpack-luau because they aren't UTF-8 encoded.
|
|
||||||
if file_name.ends_with(".png") {
|
|
||||||
continue;
|
|
||||||
}
|
|
||||||
|
|
||||||
let child_snapshot = snapshot_from_fs_path(&entry.path())?;
|
let child_snapshot = snapshot_from_fs_path(&entry.path())?;
|
||||||
children.push((file_name, child_snapshot));
|
children.push((file_name, child_snapshot));
|
||||||
}
|
}
|
||||||
@@ -75,7 +70,6 @@ fn main() -> Result<(), anyhow::Error> {
|
|||||||
"src" => snapshot_from_fs_path(&plugin_dir.join("src"))?,
|
"src" => snapshot_from_fs_path(&plugin_dir.join("src"))?,
|
||||||
"Packages" => snapshot_from_fs_path(&plugin_dir.join("Packages"))?,
|
"Packages" => snapshot_from_fs_path(&plugin_dir.join("Packages"))?,
|
||||||
"Version.txt" => snapshot_from_fs_path(&plugin_dir.join("Version.txt"))?,
|
"Version.txt" => snapshot_from_fs_path(&plugin_dir.join("Version.txt"))?,
|
||||||
"UploadDetails.json" => snapshot_from_fs_path(&plugin_dir.join("UploadDetails.json"))?,
|
|
||||||
}),
|
}),
|
||||||
});
|
});
|
||||||
|
|
||||||
|
|||||||
@@ -22,9 +22,6 @@
|
|||||||
},
|
},
|
||||||
"Version": {
|
"Version": {
|
||||||
"$path": "plugin/Version.txt"
|
"$path": "plugin/Version.txt"
|
||||||
},
|
|
||||||
"UploadDetails": {
|
|
||||||
"$path": "plugin/UploadDetails.json"
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
Submodule plugin/Packages/msgpack-luau deleted from 40f67fc0f6
@@ -1,7 +0,0 @@
|
|||||||
{
|
|
||||||
"assetId": 13916111004,
|
|
||||||
"name": "Rojo",
|
|
||||||
"description": "The plugin portion of Rojo, a tool to enable professional tooling for Roblox developers.",
|
|
||||||
"creatorId": 32644114,
|
|
||||||
"creatorType": "Group"
|
|
||||||
}
|
|
||||||
@@ -1,7 +1,5 @@
|
|||||||
local HttpService = game:GetService("HttpService")
|
local HttpService = game:GetService("HttpService")
|
||||||
|
|
||||||
local msgpack = require(script.Parent.Parent.msgpack)
|
|
||||||
|
|
||||||
local stringTemplate = [[
|
local stringTemplate = [[
|
||||||
Http.Response {
|
Http.Response {
|
||||||
code: %d
|
code: %d
|
||||||
@@ -33,8 +31,4 @@ function Response:json()
|
|||||||
return HttpService:JSONDecode(self.body)
|
return HttpService:JSONDecode(self.body)
|
||||||
end
|
end
|
||||||
|
|
||||||
function Response:msgpack()
|
|
||||||
return msgpack.decode(self.body)
|
|
||||||
end
|
|
||||||
|
|
||||||
return Response
|
return Response
|
||||||
|
|||||||
@@ -1,8 +1,7 @@
|
|||||||
local HttpService = game:GetService("HttpService")
|
local HttpService = game:GetService("HttpService")
|
||||||
|
|
||||||
local Log = require(script.Parent.Log)
|
|
||||||
local msgpack = require(script.Parent.msgpack)
|
|
||||||
local Promise = require(script.Parent.Promise)
|
local Promise = require(script.Parent.Promise)
|
||||||
|
local Log = require(script.Parent.Log)
|
||||||
|
|
||||||
local HttpError = require(script.Error)
|
local HttpError = require(script.Error)
|
||||||
local HttpResponse = require(script.Response)
|
local HttpResponse = require(script.Response)
|
||||||
@@ -69,12 +68,4 @@ function Http.jsonDecode(source)
|
|||||||
return HttpService:JSONDecode(source)
|
return HttpService:JSONDecode(source)
|
||||||
end
|
end
|
||||||
|
|
||||||
function Http.msgpackEncode(object)
|
|
||||||
return msgpack.encode(object)
|
|
||||||
end
|
|
||||||
|
|
||||||
function Http.msgpackDecode(source)
|
|
||||||
return msgpack.decode(source)
|
|
||||||
end
|
|
||||||
|
|
||||||
return Http
|
return Http
|
||||||
|
|||||||
@@ -145,7 +145,7 @@ function ApiContext:connect()
|
|||||||
|
|
||||||
return Http.get(url)
|
return Http.get(url)
|
||||||
:andThen(rejectFailedRequests)
|
:andThen(rejectFailedRequests)
|
||||||
:andThen(Http.Response.msgpack)
|
:andThen(Http.Response.json)
|
||||||
:andThen(rejectWrongProtocolVersion)
|
:andThen(rejectWrongProtocolVersion)
|
||||||
:andThen(function(body)
|
:andThen(function(body)
|
||||||
assert(validateApiInfo(body))
|
assert(validateApiInfo(body))
|
||||||
@@ -163,7 +163,7 @@ end
|
|||||||
function ApiContext:read(ids)
|
function ApiContext:read(ids)
|
||||||
local url = ("%s/api/read/%s"):format(self.__baseUrl, table.concat(ids, ","))
|
local url = ("%s/api/read/%s"):format(self.__baseUrl, table.concat(ids, ","))
|
||||||
|
|
||||||
return Http.get(url):andThen(rejectFailedRequests):andThen(Http.Response.msgpack):andThen(function(body)
|
return Http.get(url):andThen(rejectFailedRequests):andThen(Http.Response.json):andThen(function(body)
|
||||||
if body.sessionId ~= self.__sessionId then
|
if body.sessionId ~= self.__sessionId then
|
||||||
return Promise.reject("Server changed ID")
|
return Promise.reject("Server changed ID")
|
||||||
end
|
end
|
||||||
@@ -191,9 +191,9 @@ function ApiContext:write(patch)
|
|||||||
table.insert(updated, fixedUpdate)
|
table.insert(updated, fixedUpdate)
|
||||||
end
|
end
|
||||||
|
|
||||||
-- Only add the 'added' field if the table is non-empty, or else the msgpack
|
-- Only add the 'added' field if the table is non-empty, or else Roblox's
|
||||||
-- encode implementation will turn the table into an array instead of a map,
|
-- JSON implementation will turn the table into an array instead of an
|
||||||
-- causing API validation to fail.
|
-- object, causing API validation to fail.
|
||||||
local added
|
local added
|
||||||
if next(patch.added) ~= nil then
|
if next(patch.added) ~= nil then
|
||||||
added = patch.added
|
added = patch.added
|
||||||
@@ -206,16 +206,13 @@ function ApiContext:write(patch)
|
|||||||
added = added,
|
added = added,
|
||||||
}
|
}
|
||||||
|
|
||||||
body = Http.msgpackEncode(body)
|
body = Http.jsonEncode(body)
|
||||||
|
|
||||||
return Http.post(url, body)
|
return Http.post(url, body):andThen(rejectFailedRequests):andThen(Http.Response.json):andThen(function(responseBody)
|
||||||
:andThen(rejectFailedRequests)
|
Log.info("Write response: {:?}", responseBody)
|
||||||
:andThen(Http.Response.msgpack)
|
|
||||||
:andThen(function(responseBody)
|
|
||||||
Log.info("Write response: {:?}", responseBody)
|
|
||||||
|
|
||||||
return responseBody
|
return responseBody
|
||||||
end)
|
end)
|
||||||
end
|
end
|
||||||
|
|
||||||
function ApiContext:connectWebSocket(packetHandlers)
|
function ApiContext:connectWebSocket(packetHandlers)
|
||||||
@@ -237,7 +234,7 @@ function ApiContext:connectWebSocket(packetHandlers)
|
|||||||
local closed, errored, received
|
local closed, errored, received
|
||||||
|
|
||||||
received = self.__wsClient.MessageReceived:Connect(function(msg)
|
received = self.__wsClient.MessageReceived:Connect(function(msg)
|
||||||
local data = Http.msgpackDecode(msg)
|
local data = Http.jsonDecode(msg)
|
||||||
if data.sessionId ~= self.__sessionId then
|
if data.sessionId ~= self.__sessionId then
|
||||||
Log.warn("Received message with wrong session ID; ignoring")
|
Log.warn("Received message with wrong session ID; ignoring")
|
||||||
return
|
return
|
||||||
@@ -283,7 +280,7 @@ end
|
|||||||
function ApiContext:open(id)
|
function ApiContext:open(id)
|
||||||
local url = ("%s/api/open/%s"):format(self.__baseUrl, id)
|
local url = ("%s/api/open/%s"):format(self.__baseUrl, id)
|
||||||
|
|
||||||
return Http.post(url, ""):andThen(rejectFailedRequests):andThen(Http.Response.msgpack):andThen(function(body)
|
return Http.post(url, ""):andThen(rejectFailedRequests):andThen(Http.Response.json):andThen(function(body)
|
||||||
if body.sessionId ~= self.__sessionId then
|
if body.sessionId ~= self.__sessionId then
|
||||||
return Promise.reject("Server changed ID")
|
return Promise.reject("Server changed ID")
|
||||||
end
|
end
|
||||||
@@ -294,11 +291,11 @@ end
|
|||||||
|
|
||||||
function ApiContext:serialize(ids: { string })
|
function ApiContext:serialize(ids: { string })
|
||||||
local url = ("%s/api/serialize"):format(self.__baseUrl)
|
local url = ("%s/api/serialize"):format(self.__baseUrl)
|
||||||
local request_body = Http.msgpackEncode({ sessionId = self.__sessionId, ids = ids })
|
local request_body = Http.jsonEncode({ sessionId = self.__sessionId, ids = ids })
|
||||||
|
|
||||||
return Http.post(url, request_body)
|
return Http.post(url, request_body)
|
||||||
:andThen(rejectFailedRequests)
|
:andThen(rejectFailedRequests)
|
||||||
:andThen(Http.Response.msgpack)
|
:andThen(Http.Response.json)
|
||||||
:andThen(function(response_body)
|
:andThen(function(response_body)
|
||||||
if response_body.sessionId ~= self.__sessionId then
|
if response_body.sessionId ~= self.__sessionId then
|
||||||
return Promise.reject("Server changed ID")
|
return Promise.reject("Server changed ID")
|
||||||
@@ -312,11 +309,11 @@ end
|
|||||||
|
|
||||||
function ApiContext:refPatch(ids: { string })
|
function ApiContext:refPatch(ids: { string })
|
||||||
local url = ("%s/api/ref-patch"):format(self.__baseUrl)
|
local url = ("%s/api/ref-patch"):format(self.__baseUrl)
|
||||||
local request_body = Http.msgpackEncode({ sessionId = self.__sessionId, ids = ids })
|
local request_body = Http.jsonEncode({ sessionId = self.__sessionId, ids = ids })
|
||||||
|
|
||||||
return Http.post(url, request_body)
|
return Http.post(url, request_body)
|
||||||
:andThen(rejectFailedRequests)
|
:andThen(rejectFailedRequests)
|
||||||
:andThen(Http.Response.msgpack)
|
:andThen(Http.Response.json)
|
||||||
:andThen(function(response_body)
|
:andThen(function(response_body)
|
||||||
if response_body.sessionId ~= self.__sessionId then
|
if response_body.sessionId ~= self.__sessionId then
|
||||||
return Promise.reject("Server changed ID")
|
return Promise.reject("Server changed ID")
|
||||||
|
|||||||
@@ -19,15 +19,9 @@ local FullscreenNotification = Roact.Component:extend("FullscreeFullscreenNotifi
|
|||||||
function FullscreenNotification:init()
|
function FullscreenNotification:init()
|
||||||
self.transparency, self.setTransparency = Roact.createBinding(0)
|
self.transparency, self.setTransparency = Roact.createBinding(0)
|
||||||
self.lifetime = self.props.timeout
|
self.lifetime = self.props.timeout
|
||||||
self.dismissed = false
|
|
||||||
end
|
end
|
||||||
|
|
||||||
function FullscreenNotification:dismiss()
|
function FullscreenNotification:dismiss()
|
||||||
if self.dismissed then
|
|
||||||
return
|
|
||||||
end
|
|
||||||
self.dismissed = true
|
|
||||||
|
|
||||||
if self.props.onClose then
|
if self.props.onClose then
|
||||||
self.props.onClose()
|
self.props.onClose()
|
||||||
end
|
end
|
||||||
@@ -65,7 +59,7 @@ function FullscreenNotification:didMount()
|
|||||||
end
|
end
|
||||||
|
|
||||||
function FullscreenNotification:willUnmount()
|
function FullscreenNotification:willUnmount()
|
||||||
if self.timeout and coroutine.status(self.timeout) == "suspended" then
|
if self.timeout and coroutine.status(self.timeout) ~= "dead" then
|
||||||
task.cancel(self.timeout)
|
task.cancel(self.timeout)
|
||||||
end
|
end
|
||||||
end
|
end
|
||||||
|
|||||||
@@ -25,7 +25,6 @@ function Notification:init()
|
|||||||
self.binding = bindingUtil.fromMotor(self.motor)
|
self.binding = bindingUtil.fromMotor(self.motor)
|
||||||
|
|
||||||
self.lifetime = self.props.timeout
|
self.lifetime = self.props.timeout
|
||||||
self.dismissed = false
|
|
||||||
|
|
||||||
self.motor:onStep(function(value)
|
self.motor:onStep(function(value)
|
||||||
if value <= 0 and self.props.onClose then
|
if value <= 0 and self.props.onClose then
|
||||||
@@ -35,11 +34,6 @@ function Notification:init()
|
|||||||
end
|
end
|
||||||
|
|
||||||
function Notification:dismiss()
|
function Notification:dismiss()
|
||||||
if self.dismissed then
|
|
||||||
return
|
|
||||||
end
|
|
||||||
self.dismissed = true
|
|
||||||
|
|
||||||
self.motor:setGoal(Flipper.Spring.new(0, {
|
self.motor:setGoal(Flipper.Spring.new(0, {
|
||||||
frequency = 5,
|
frequency = 5,
|
||||||
dampingRatio = 1,
|
dampingRatio = 1,
|
||||||
@@ -81,7 +75,7 @@ function Notification:didMount()
|
|||||||
end
|
end
|
||||||
|
|
||||||
function Notification:willUnmount()
|
function Notification:willUnmount()
|
||||||
if self.timeout and coroutine.status(self.timeout) == "suspended" then
|
if self.timeout and coroutine.status(self.timeout) ~= "dead" then
|
||||||
task.cancel(self.timeout)
|
task.cancel(self.timeout)
|
||||||
end
|
end
|
||||||
end
|
end
|
||||||
|
|||||||
@@ -301,19 +301,6 @@ function App:setPriorSyncInfo(host: string, port: string, projectName: string)
|
|||||||
Settings:set("priorEndpoints", priorSyncInfos)
|
Settings:set("priorEndpoints", priorSyncInfos)
|
||||||
end
|
end
|
||||||
|
|
||||||
function App:forgetPriorSyncInfo()
|
|
||||||
local priorSyncInfos = Settings:get("priorEndpoints")
|
|
||||||
if not priorSyncInfos then
|
|
||||||
priorSyncInfos = {}
|
|
||||||
end
|
|
||||||
|
|
||||||
local id = tostring(game.PlaceId)
|
|
||||||
priorSyncInfos[id] = nil
|
|
||||||
Log.trace("Erased last used endpoint for {}", game.PlaceId)
|
|
||||||
|
|
||||||
Settings:set("priorEndpoints", priorSyncInfos)
|
|
||||||
end
|
|
||||||
|
|
||||||
function App:getHostAndPort()
|
function App:getHostAndPort()
|
||||||
local host = self.host:getValue()
|
local host = self.host:getValue()
|
||||||
local port = self.port:getValue()
|
local port = self.port:getValue()
|
||||||
@@ -448,8 +435,7 @@ function App:checkSyncReminder()
|
|||||||
self:findActiveServer()
|
self:findActiveServer()
|
||||||
:andThen(function(serverInfo, host, port)
|
:andThen(function(serverInfo, host, port)
|
||||||
self:sendSyncReminder(
|
self:sendSyncReminder(
|
||||||
`Project '{serverInfo.projectName}' is serving at {host}:{port}.\nWould you like to connect?`,
|
`Project '{serverInfo.projectName}' is serving at {host}:{port}.\nWould you like to connect?`
|
||||||
{ "Connect", "Dismiss" }
|
|
||||||
)
|
)
|
||||||
end)
|
end)
|
||||||
:catch(function()
|
:catch(function()
|
||||||
@@ -460,8 +446,7 @@ function App:checkSyncReminder()
|
|||||||
|
|
||||||
local timeSinceSync = timeUtil.elapsedToText(os.time() - priorSyncInfo.timestamp)
|
local timeSinceSync = timeUtil.elapsedToText(os.time() - priorSyncInfo.timestamp)
|
||||||
self:sendSyncReminder(
|
self:sendSyncReminder(
|
||||||
`You synced project '{priorSyncInfo.projectName}' to this place {timeSinceSync}.\nDid you mean to run 'rojo serve' and then connect?`,
|
`You synced project '{priorSyncInfo.projectName}' to this place {timeSinceSync}.\nDid you mean to run 'rojo serve' and then connect?`
|
||||||
{ "Connect", "Forget", "Dismiss" }
|
|
||||||
)
|
)
|
||||||
end
|
end
|
||||||
end)
|
end)
|
||||||
@@ -501,16 +486,12 @@ function App:stopSyncReminderPolling()
|
|||||||
end
|
end
|
||||||
end
|
end
|
||||||
|
|
||||||
function App:sendSyncReminder(message: string, shownActions: { string })
|
function App:sendSyncReminder(message: string)
|
||||||
local syncReminderMode = Settings:get("syncReminderMode")
|
local syncReminderMode = Settings:get("syncReminderMode")
|
||||||
if syncReminderMode == "None" then
|
if syncReminderMode == "None" then
|
||||||
return
|
return
|
||||||
end
|
end
|
||||||
|
|
||||||
local connectIndex = table.find(shownActions, "Connect")
|
|
||||||
local forgetIndex = table.find(shownActions, "Forget")
|
|
||||||
local dismissIndex = table.find(shownActions, "Dismiss")
|
|
||||||
|
|
||||||
self.dismissSyncReminder = self:addNotification({
|
self.dismissSyncReminder = self:addNotification({
|
||||||
text = message,
|
text = message,
|
||||||
timeout = 120,
|
timeout = 120,
|
||||||
@@ -519,39 +500,24 @@ function App:sendSyncReminder(message: string, shownActions: { string })
|
|||||||
self.dismissSyncReminder = nil
|
self.dismissSyncReminder = nil
|
||||||
end,
|
end,
|
||||||
actions = {
|
actions = {
|
||||||
Connect = if connectIndex
|
Connect = {
|
||||||
then {
|
text = "Connect",
|
||||||
text = "Connect",
|
style = "Solid",
|
||||||
style = "Solid",
|
layoutOrder = 1,
|
||||||
layoutOrder = connectIndex,
|
onClick = function()
|
||||||
onClick = function()
|
self:startSession()
|
||||||
self:startSession()
|
end,
|
||||||
end,
|
},
|
||||||
}
|
Dismiss = {
|
||||||
else nil,
|
text = "Dismiss",
|
||||||
Forget = if forgetIndex
|
style = "Bordered",
|
||||||
then {
|
layoutOrder = 2,
|
||||||
text = "Forget",
|
onClick = function()
|
||||||
style = "Bordered",
|
-- If the user dismisses the reminder,
|
||||||
layoutOrder = forgetIndex,
|
-- then we don't need to remind them again
|
||||||
onClick = function()
|
self:stopSyncReminderPolling()
|
||||||
-- The user doesn't want to be reminded again about this sync
|
end,
|
||||||
self:forgetPriorSyncInfo()
|
},
|
||||||
end,
|
|
||||||
}
|
|
||||||
else nil,
|
|
||||||
Dismiss = if dismissIndex
|
|
||||||
then {
|
|
||||||
text = "Dismiss",
|
|
||||||
style = "Bordered",
|
|
||||||
layoutOrder = dismissIndex,
|
|
||||||
onClick = function()
|
|
||||||
-- If the user dismisses the reminder,
|
|
||||||
-- then we don't need to remind them again
|
|
||||||
self:stopSyncReminderPolling()
|
|
||||||
end,
|
|
||||||
}
|
|
||||||
else nil,
|
|
||||||
},
|
},
|
||||||
})
|
})
|
||||||
end
|
end
|
||||||
|
|||||||
@@ -54,10 +54,6 @@ local function trueEquals(a, b): boolean
|
|||||||
end
|
end
|
||||||
return true
|
return true
|
||||||
|
|
||||||
-- For NaN, check if both values are not equal to themselves
|
|
||||||
elseif a ~= a and b ~= b then
|
|
||||||
return true
|
|
||||||
|
|
||||||
-- For numbers, compare with epsilon of 0.0001 to avoid floating point inequality
|
-- For numbers, compare with epsilon of 0.0001 to avoid floating point inequality
|
||||||
elseif typeA == "number" and typeB == "number" then
|
elseif typeA == "number" and typeB == "number" then
|
||||||
return fuzzyEq(a, b, 0.0001)
|
return fuzzyEq(a, b, 0.0001)
|
||||||
|
|||||||
@@ -1,16 +0,0 @@
|
|||||||
---
|
|
||||||
source: tests/tests/build.rs
|
|
||||||
expression: contents
|
|
||||||
---
|
|
||||||
<roblox version="4">
|
|
||||||
<Item class="Folder" referent="0">
|
|
||||||
<Properties>
|
|
||||||
<string name="Name">json_model_legacy_name</string>
|
|
||||||
</Properties>
|
|
||||||
<Item class="Folder" referent="1">
|
|
||||||
<Properties>
|
|
||||||
<string name="Name">Expected Name</string>
|
|
||||||
</Properties>
|
|
||||||
</Item>
|
|
||||||
</Item>
|
|
||||||
</roblox>
|
|
||||||
@@ -0,0 +1,23 @@
|
|||||||
|
---
|
||||||
|
source: tests/tests/build.rs
|
||||||
|
assertion_line: 109
|
||||||
|
expression: contents
|
||||||
|
---
|
||||||
|
<roblox version="4">
|
||||||
|
<Item class="DataModel" referent="0">
|
||||||
|
<Properties>
|
||||||
|
<string name="Name">model_json_name_input</string>
|
||||||
|
</Properties>
|
||||||
|
<Item class="Workspace" referent="1">
|
||||||
|
<Properties>
|
||||||
|
<string name="Name">Workspace</string>
|
||||||
|
<bool name="NeedsPivotMigration">false</bool>
|
||||||
|
</Properties>
|
||||||
|
<Item class="StringValue" referent="2">
|
||||||
|
<Properties>
|
||||||
|
<string name="Name">/Bar</string>
|
||||||
|
</Properties>
|
||||||
|
</Item>
|
||||||
|
</Item>
|
||||||
|
</Item>
|
||||||
|
</roblox>
|
||||||
@@ -0,0 +1,20 @@
|
|||||||
|
---
|
||||||
|
source: tests/tests/build.rs
|
||||||
|
assertion_line: 108
|
||||||
|
expression: contents
|
||||||
|
---
|
||||||
|
<roblox version="4">
|
||||||
|
<Item class="Folder" referent="0">
|
||||||
|
<Properties>
|
||||||
|
<string name="Name">slugified_name_roundtrip</string>
|
||||||
|
</Properties>
|
||||||
|
<Item class="Script" referent="1">
|
||||||
|
<Properties>
|
||||||
|
<string name="Name">/Script</string>
|
||||||
|
<token name="RunContext">0</token>
|
||||||
|
<string name="Source"><![CDATA[print("Hello world!")
|
||||||
|
]]></string>
|
||||||
|
</Properties>
|
||||||
|
</Item>
|
||||||
|
</Item>
|
||||||
|
</roblox>
|
||||||
@@ -1,6 +0,0 @@
|
|||||||
{
|
|
||||||
"name": "json_model_legacy_name",
|
|
||||||
"tree": {
|
|
||||||
"$path": "folder"
|
|
||||||
}
|
|
||||||
}
|
|
||||||
@@ -1,4 +0,0 @@
|
|||||||
{
|
|
||||||
"Name": "Overridden Name",
|
|
||||||
"ClassName": "Folder"
|
|
||||||
}
|
|
||||||
@@ -0,0 +1,11 @@
|
|||||||
|
{
|
||||||
|
"name": "model_json_name_input",
|
||||||
|
"tree": {
|
||||||
|
"$className": "DataModel",
|
||||||
|
"Workspace": {
|
||||||
|
"$className": "Workspace",
|
||||||
|
"$path": "src"
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
@@ -0,0 +1,5 @@
|
|||||||
|
{
|
||||||
|
"name": "/Bar",
|
||||||
|
"className": "StringValue"
|
||||||
|
}
|
||||||
|
|
||||||
@@ -0,0 +1,4 @@
|
|||||||
|
{
|
||||||
|
"name": "/Script"
|
||||||
|
}
|
||||||
|
|
||||||
@@ -0,0 +1,2 @@
|
|||||||
|
print("Hello world!")
|
||||||
|
|
||||||
@@ -0,0 +1,6 @@
|
|||||||
|
{
|
||||||
|
"name": "slugified_name_roundtrip",
|
||||||
|
"tree": {
|
||||||
|
"$path": "src"
|
||||||
|
}
|
||||||
|
}
|
||||||
@@ -0,0 +1,3 @@
|
|||||||
|
{
|
||||||
|
"name": "/Script"
|
||||||
|
}
|
||||||
@@ -0,0 +1 @@
|
|||||||
|
print("Hello world!")
|
||||||
@@ -0,0 +1,6 @@
|
|||||||
|
---
|
||||||
|
source: tests/rojo_test/syncback_util.rs
|
||||||
|
assertion_line: 101
|
||||||
|
expression: "String::from_utf8_lossy(&output.stdout)"
|
||||||
|
---
|
||||||
|
|
||||||
@@ -0,0 +1,13 @@
|
|||||||
|
---
|
||||||
|
source: tests/rojo_test/syncback_util.rs
|
||||||
|
assertion_line: 101
|
||||||
|
expression: "String::from_utf8_lossy(&output.stdout)"
|
||||||
|
---
|
||||||
|
Writing default.project.json
|
||||||
|
Writing src/Camera.rbxm
|
||||||
|
Writing src/Terrain.rbxm
|
||||||
|
Writing src/_Folder/init.meta.json
|
||||||
|
Writing src/_Script.meta.json
|
||||||
|
Writing src/_Script.server.luau
|
||||||
|
Writing src
|
||||||
|
Writing src/_Folder
|
||||||
@@ -0,0 +1,9 @@
|
|||||||
|
---
|
||||||
|
source: tests/tests/syncback.rs
|
||||||
|
assertion_line: 31
|
||||||
|
expression: src/foo.model.json
|
||||||
|
---
|
||||||
|
{
|
||||||
|
"name": "/Bar",
|
||||||
|
"className": "StringValue"
|
||||||
|
}
|
||||||
@@ -0,0 +1,8 @@
|
|||||||
|
---
|
||||||
|
source: tests/tests/syncback.rs
|
||||||
|
assertion_line: 31
|
||||||
|
expression: src/_Folder.model.json
|
||||||
|
---
|
||||||
|
{
|
||||||
|
"className": "Folder"
|
||||||
|
}
|
||||||
@@ -0,0 +1,8 @@
|
|||||||
|
---
|
||||||
|
source: tests/tests/syncback.rs
|
||||||
|
assertion_line: 31
|
||||||
|
expression: src/_Folder/init.meta.json
|
||||||
|
---
|
||||||
|
{
|
||||||
|
"name": "/Folder"
|
||||||
|
}
|
||||||
@@ -0,0 +1,8 @@
|
|||||||
|
---
|
||||||
|
source: tests/tests/syncback.rs
|
||||||
|
assertion_line: 31
|
||||||
|
expression: src/_Script.meta.json
|
||||||
|
---
|
||||||
|
{
|
||||||
|
"name": "/Script"
|
||||||
|
}
|
||||||
@@ -0,0 +1,6 @@
|
|||||||
|
---
|
||||||
|
source: tests/tests/syncback.rs
|
||||||
|
assertion_line: 31
|
||||||
|
expression: src/_Script.server.luau
|
||||||
|
---
|
||||||
|
print("Hello world!")
|
||||||
@@ -0,0 +1,8 @@
|
|||||||
|
---
|
||||||
|
source: tests/tests/syncback.rs
|
||||||
|
assertion_line: 31
|
||||||
|
expression: src/_Script/init.meta.json
|
||||||
|
---
|
||||||
|
{
|
||||||
|
"name": "/Script"
|
||||||
|
}
|
||||||
@@ -0,0 +1,6 @@
|
|||||||
|
---
|
||||||
|
source: tests/tests/syncback.rs
|
||||||
|
assertion_line: 31
|
||||||
|
expression: src/_Script/init.server.luau
|
||||||
|
---
|
||||||
|
print("Hello world!")
|
||||||
@@ -0,0 +1,11 @@
|
|||||||
|
{
|
||||||
|
"name": "model_json_name",
|
||||||
|
"tree": {
|
||||||
|
"$className": "DataModel",
|
||||||
|
"Workspace": {
|
||||||
|
"$className": "Workspace",
|
||||||
|
"$path": "src"
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
@@ -0,0 +1,5 @@
|
|||||||
|
{
|
||||||
|
"name": "/Bar",
|
||||||
|
"className": "StringValue"
|
||||||
|
}
|
||||||
|
|
||||||
BIN
rojo-test/syncback-tests/model_json_name/input.rbxl
Normal file
BIN
rojo-test/syncback-tests/model_json_name/input.rbxl
Normal file
Binary file not shown.
@@ -0,0 +1,10 @@
|
|||||||
|
{
|
||||||
|
"name": "slugified_name",
|
||||||
|
"tree": {
|
||||||
|
"$className": "DataModel",
|
||||||
|
"Workspace": {
|
||||||
|
"$className": "Workspace",
|
||||||
|
"$path": "src"
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
BIN
rojo-test/syncback-tests/slugified_name/input.rbxl
Normal file
BIN
rojo-test/syncback-tests/slugified_name/input.rbxl
Normal file
Binary file not shown.
@@ -3,4 +3,3 @@ rojo = "rojo-rbx/rojo@7.5.1"
|
|||||||
selene = "Kampfkarren/selene@0.29.0"
|
selene = "Kampfkarren/selene@0.29.0"
|
||||||
stylua = "JohnnyMorganz/stylua@2.1.0"
|
stylua = "JohnnyMorganz/stylua@2.1.0"
|
||||||
run-in-roblox = "rojo-rbx/run-in-roblox@0.3.0"
|
run-in-roblox = "rojo-rbx/run-in-roblox@0.3.0"
|
||||||
lune = "lune-org/lune@0.10.4"
|
|
||||||
|
|||||||
@@ -98,5 +98,5 @@ fn uninstall_plugin() -> anyhow::Result<()> {
|
|||||||
|
|
||||||
#[test]
|
#[test]
|
||||||
fn plugin_initialize() {
|
fn plugin_initialize() {
|
||||||
let _ = initialize_plugin().unwrap();
|
assert!(initialize_plugin().is_ok())
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -1,35 +0,0 @@
|
|||||||
---
|
|
||||||
source: src/cli/sourcemap.rs
|
|
||||||
expression: sourcemap_contents
|
|
||||||
---
|
|
||||||
{
|
|
||||||
"name": "default",
|
|
||||||
"className": "DataModel",
|
|
||||||
"filePaths": "[...1 path omitted...]",
|
|
||||||
"children": [
|
|
||||||
{
|
|
||||||
"name": "ReplicatedStorage",
|
|
||||||
"className": "ReplicatedStorage",
|
|
||||||
"children": [
|
|
||||||
{
|
|
||||||
"name": "Project",
|
|
||||||
"className": "ModuleScript",
|
|
||||||
"filePaths": "[...1 path omitted...]",
|
|
||||||
"children": [
|
|
||||||
{
|
|
||||||
"name": "Module",
|
|
||||||
"className": "Folder",
|
|
||||||
"children": [
|
|
||||||
{
|
|
||||||
"name": "module",
|
|
||||||
"className": "ModuleScript",
|
|
||||||
"filePaths": "[...1 path omitted...]"
|
|
||||||
}
|
|
||||||
]
|
|
||||||
}
|
|
||||||
]
|
|
||||||
}
|
|
||||||
]
|
|
||||||
}
|
|
||||||
]
|
|
||||||
}
|
|
||||||
@@ -1,41 +0,0 @@
|
|||||||
---
|
|
||||||
source: src/cli/sourcemap.rs
|
|
||||||
expression: sourcemap_contents
|
|
||||||
---
|
|
||||||
{
|
|
||||||
"name": "default",
|
|
||||||
"className": "DataModel",
|
|
||||||
"filePaths": [
|
|
||||||
"default.project.json"
|
|
||||||
],
|
|
||||||
"children": [
|
|
||||||
{
|
|
||||||
"name": "ReplicatedStorage",
|
|
||||||
"className": "ReplicatedStorage",
|
|
||||||
"children": [
|
|
||||||
{
|
|
||||||
"name": "Project",
|
|
||||||
"className": "ModuleScript",
|
|
||||||
"filePaths": [
|
|
||||||
"src/init.luau"
|
|
||||||
],
|
|
||||||
"children": [
|
|
||||||
{
|
|
||||||
"name": "Module",
|
|
||||||
"className": "Folder",
|
|
||||||
"children": [
|
|
||||||
{
|
|
||||||
"name": "module",
|
|
||||||
"className": "ModuleScript",
|
|
||||||
"filePaths": [
|
|
||||||
"../module/module.luau"
|
|
||||||
]
|
|
||||||
}
|
|
||||||
]
|
|
||||||
}
|
|
||||||
]
|
|
||||||
}
|
|
||||||
]
|
|
||||||
}
|
|
||||||
]
|
|
||||||
}
|
|
||||||
@@ -10,7 +10,7 @@ use fs_err::File;
|
|||||||
use memofs::Vfs;
|
use memofs::Vfs;
|
||||||
use rayon::prelude::*;
|
use rayon::prelude::*;
|
||||||
use rbx_dom_weak::{types::Ref, Ustr};
|
use rbx_dom_weak::{types::Ref, Ustr};
|
||||||
use serde::{Deserialize, Serialize};
|
use serde::Serialize;
|
||||||
use tokio::runtime::Runtime;
|
use tokio::runtime::Runtime;
|
||||||
|
|
||||||
use crate::{
|
use crate::{
|
||||||
@@ -24,20 +24,19 @@ const PATH_STRIP_FAILED_ERR: &str = "Failed to create relative paths for project
|
|||||||
const ABSOLUTE_PATH_FAILED_ERR: &str = "Failed to turn relative path into absolute path!";
|
const ABSOLUTE_PATH_FAILED_ERR: &str = "Failed to turn relative path into absolute path!";
|
||||||
|
|
||||||
/// Representation of a node in the generated sourcemap tree.
|
/// Representation of a node in the generated sourcemap tree.
|
||||||
#[derive(Serialize, Deserialize)]
|
#[derive(Serialize)]
|
||||||
#[serde(rename_all = "camelCase")]
|
#[serde(rename_all = "camelCase")]
|
||||||
struct SourcemapNode<'a> {
|
struct SourcemapNode<'a> {
|
||||||
name: &'a str,
|
name: &'a str,
|
||||||
class_name: Ustr,
|
class_name: Ustr,
|
||||||
|
|
||||||
#[serde(
|
#[serde(
|
||||||
default,
|
|
||||||
skip_serializing_if = "Vec::is_empty",
|
skip_serializing_if = "Vec::is_empty",
|
||||||
serialize_with = "crate::path_serializer::serialize_vec_absolute"
|
serialize_with = "crate::path_serializer::serialize_vec_absolute"
|
||||||
)]
|
)]
|
||||||
file_paths: Vec<Cow<'a, Path>>,
|
file_paths: Vec<Cow<'a, Path>>,
|
||||||
|
|
||||||
#[serde(default, skip_serializing_if = "Vec::is_empty")]
|
#[serde(skip_serializing_if = "Vec::is_empty")]
|
||||||
children: Vec<SourcemapNode<'a>>,
|
children: Vec<SourcemapNode<'a>>,
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -71,13 +70,12 @@ pub struct SourcemapCommand {
|
|||||||
|
|
||||||
impl SourcemapCommand {
|
impl SourcemapCommand {
|
||||||
pub fn run(self) -> anyhow::Result<()> {
|
pub fn run(self) -> anyhow::Result<()> {
|
||||||
let project_path = fs_err::canonicalize(resolve_path(&self.project))?;
|
let project_path = resolve_path(&self.project);
|
||||||
|
|
||||||
log::trace!("Constructing filesystem with StdBackend");
|
log::trace!("Constructing in-memory filesystem");
|
||||||
let vfs = Vfs::new_default();
|
let vfs = Vfs::new_default();
|
||||||
vfs.set_watch_enabled(self.watch);
|
vfs.set_watch_enabled(self.watch);
|
||||||
|
|
||||||
log::trace!("Setting up session for sourcemap generation");
|
|
||||||
let session = ServeSession::new(vfs, project_path)?;
|
let session = ServeSession::new(vfs, project_path)?;
|
||||||
let mut cursor = session.message_queue().cursor();
|
let mut cursor = session.message_queue().cursor();
|
||||||
|
|
||||||
@@ -89,17 +87,14 @@ impl SourcemapCommand {
|
|||||||
|
|
||||||
// Pre-build a rayon threadpool with a low number of threads to avoid
|
// Pre-build a rayon threadpool with a low number of threads to avoid
|
||||||
// dynamic creation overhead on systems with a high number of cpus.
|
// dynamic creation overhead on systems with a high number of cpus.
|
||||||
log::trace!("Setting rayon global threadpool");
|
|
||||||
rayon::ThreadPoolBuilder::new()
|
rayon::ThreadPoolBuilder::new()
|
||||||
.num_threads(num_cpus::get().min(6))
|
.num_threads(num_cpus::get().min(6))
|
||||||
.build_global()
|
.build_global()
|
||||||
.ok();
|
.unwrap();
|
||||||
|
|
||||||
log::trace!("Writing initial sourcemap");
|
|
||||||
write_sourcemap(&session, self.output.as_deref(), filter, self.absolute)?;
|
write_sourcemap(&session, self.output.as_deref(), filter, self.absolute)?;
|
||||||
|
|
||||||
if self.watch {
|
if self.watch {
|
||||||
log::trace!("Setting up runtime for watch mode");
|
|
||||||
let rt = Runtime::new().unwrap();
|
let rt = Runtime::new().unwrap();
|
||||||
|
|
||||||
loop {
|
loop {
|
||||||
@@ -213,7 +208,7 @@ fn recurse_create_node<'a>(
|
|||||||
} else {
|
} else {
|
||||||
for val in file_paths {
|
for val in file_paths {
|
||||||
output_file_paths.push(Cow::from(
|
output_file_paths.push(Cow::from(
|
||||||
pathdiff::diff_paths(val, project_dir).expect(PATH_STRIP_FAILED_ERR),
|
val.strip_prefix(project_dir).expect(PATH_STRIP_FAILED_ERR),
|
||||||
));
|
));
|
||||||
}
|
}
|
||||||
};
|
};
|
||||||
@@ -255,80 +250,3 @@ fn write_sourcemap(
|
|||||||
|
|
||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
|
|
||||||
#[cfg(test)]
|
|
||||||
mod test {
|
|
||||||
use crate::cli::sourcemap::SourcemapNode;
|
|
||||||
use crate::cli::SourcemapCommand;
|
|
||||||
use insta::internals::Content;
|
|
||||||
use std::path::Path;
|
|
||||||
|
|
||||||
#[test]
|
|
||||||
fn maps_relative_paths() {
|
|
||||||
let sourcemap_dir = tempfile::tempdir().unwrap();
|
|
||||||
let sourcemap_output = sourcemap_dir.path().join("sourcemap.json");
|
|
||||||
let project_path = fs_err::canonicalize(
|
|
||||||
Path::new(env!("CARGO_MANIFEST_DIR"))
|
|
||||||
.join("test-projects")
|
|
||||||
.join("relative_paths")
|
|
||||||
.join("project"),
|
|
||||||
)
|
|
||||||
.unwrap();
|
|
||||||
let sourcemap_command = SourcemapCommand {
|
|
||||||
project: project_path,
|
|
||||||
output: Some(sourcemap_output.clone()),
|
|
||||||
include_non_scripts: false,
|
|
||||||
watch: false,
|
|
||||||
absolute: false,
|
|
||||||
};
|
|
||||||
assert!(sourcemap_command.run().is_ok());
|
|
||||||
|
|
||||||
let raw_sourcemap_contents = fs_err::read_to_string(sourcemap_output.as_path()).unwrap();
|
|
||||||
let sourcemap_contents =
|
|
||||||
serde_json::from_str::<SourcemapNode>(&raw_sourcemap_contents).unwrap();
|
|
||||||
insta::assert_json_snapshot!(sourcemap_contents);
|
|
||||||
}
|
|
||||||
|
|
||||||
#[test]
|
|
||||||
fn maps_absolute_paths() {
|
|
||||||
let sourcemap_dir = tempfile::tempdir().unwrap();
|
|
||||||
let sourcemap_output = sourcemap_dir.path().join("sourcemap.json");
|
|
||||||
let project_path = fs_err::canonicalize(
|
|
||||||
Path::new(env!("CARGO_MANIFEST_DIR"))
|
|
||||||
.join("test-projects")
|
|
||||||
.join("relative_paths")
|
|
||||||
.join("project"),
|
|
||||||
)
|
|
||||||
.unwrap();
|
|
||||||
let sourcemap_command = SourcemapCommand {
|
|
||||||
project: project_path,
|
|
||||||
output: Some(sourcemap_output.clone()),
|
|
||||||
include_non_scripts: false,
|
|
||||||
watch: false,
|
|
||||||
absolute: true,
|
|
||||||
};
|
|
||||||
assert!(sourcemap_command.run().is_ok());
|
|
||||||
|
|
||||||
let raw_sourcemap_contents = fs_err::read_to_string(sourcemap_output.as_path()).unwrap();
|
|
||||||
let sourcemap_contents =
|
|
||||||
serde_json::from_str::<SourcemapNode>(&raw_sourcemap_contents).unwrap();
|
|
||||||
insta::assert_json_snapshot!(sourcemap_contents, {
|
|
||||||
".**.filePaths" => insta::dynamic_redaction(|mut value, _path| {
|
|
||||||
let mut paths_count = 0;
|
|
||||||
|
|
||||||
match value {
|
|
||||||
Content::Seq(ref mut vec) => {
|
|
||||||
for path in vec.iter().map(|i| i.as_str().unwrap()) {
|
|
||||||
assert_eq!(fs_err::canonicalize(path).is_ok(), true, "path was not valid");
|
|
||||||
assert_eq!(Path::new(path).is_absolute(), true, "path was not absolute");
|
|
||||||
|
|
||||||
paths_count += 1;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
_ => panic!("Expected filePaths to be a sequence"),
|
|
||||||
}
|
|
||||||
format!("[...{} path{} omitted...]", paths_count, if paths_count != 1 { "s" } else { "" } )
|
|
||||||
})
|
|
||||||
});
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|||||||
@@ -70,6 +70,12 @@ pub struct InstanceMetadata {
|
|||||||
/// A schema provided via a JSON file, if one exists. Will be `None` for
|
/// A schema provided via a JSON file, if one exists. Will be `None` for
|
||||||
/// all non-JSON middleware.
|
/// all non-JSON middleware.
|
||||||
pub schema: Option<String>,
|
pub schema: Option<String>,
|
||||||
|
|
||||||
|
/// A custom name specified via meta.json or model.json files. If present,
|
||||||
|
/// this name will be used for the instance while the filesystem name will
|
||||||
|
/// be slugified to remove illegal characters.
|
||||||
|
#[serde(skip_serializing_if = "Option::is_none")]
|
||||||
|
pub specified_name: Option<String>,
|
||||||
}
|
}
|
||||||
|
|
||||||
impl InstanceMetadata {
|
impl InstanceMetadata {
|
||||||
@@ -82,6 +88,7 @@ impl InstanceMetadata {
|
|||||||
specified_id: None,
|
specified_id: None,
|
||||||
middleware: None,
|
middleware: None,
|
||||||
schema: None,
|
schema: None,
|
||||||
|
specified_name: None,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -130,6 +137,13 @@ impl InstanceMetadata {
|
|||||||
pub fn schema(self, schema: Option<String>) -> Self {
|
pub fn schema(self, schema: Option<String>) -> Self {
|
||||||
Self { schema, ..self }
|
Self { schema, ..self }
|
||||||
}
|
}
|
||||||
|
|
||||||
|
pub fn specified_name(self, specified_name: Option<String>) -> Self {
|
||||||
|
Self {
|
||||||
|
specified_name,
|
||||||
|
..self
|
||||||
|
}
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
impl Default for InstanceMetadata {
|
impl Default for InstanceMetadata {
|
||||||
|
|||||||
@@ -8,7 +8,7 @@ use rbx_dom_weak::{
|
|||||||
ustr, HashMapExt as _, UstrMap, UstrSet,
|
ustr, HashMapExt as _, UstrMap, UstrSet,
|
||||||
};
|
};
|
||||||
|
|
||||||
use crate::{variant_eq::variant_eq, RojoRef, REF_POINTER_ATTRIBUTE_PREFIX};
|
use crate::{RojoRef, REF_POINTER_ATTRIBUTE_PREFIX};
|
||||||
|
|
||||||
use super::{
|
use super::{
|
||||||
patch::{PatchAdd, PatchSet, PatchUpdate},
|
patch::{PatchAdd, PatchSet, PatchUpdate},
|
||||||
@@ -127,7 +127,7 @@ fn compute_property_patches(
|
|||||||
|
|
||||||
match instance.properties().get(&name) {
|
match instance.properties().get(&name) {
|
||||||
Some(instance_value) => {
|
Some(instance_value) => {
|
||||||
if !variant_eq(&snapshot_value, instance_value) {
|
if &snapshot_value != instance_value {
|
||||||
changed_properties.insert(name, Some(snapshot_value));
|
changed_properties.insert(name, Some(snapshot_value));
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -35,20 +35,14 @@ pub fn snapshot_json_model(
|
|||||||
format!("File is not a valid JSON model: {}", path.display())
|
format!("File is not a valid JSON model: {}", path.display())
|
||||||
})?;
|
})?;
|
||||||
|
|
||||||
if let Some(top_level_name) = &instance.name {
|
// If the JSON has a name property, preserve it in metadata for syncback
|
||||||
let new_name = format!("{}.model.json", top_level_name);
|
let specified_name = instance.name.clone();
|
||||||
|
|
||||||
log::warn!(
|
// Use the name from JSON if present, otherwise fall back to filename-derived name
|
||||||
"Model at path {} had a top-level Name field. \
|
if instance.name.is_none() {
|
||||||
This field has been ignored since Rojo 6.0.\n\
|
instance.name = Some(name.to_owned());
|
||||||
Consider removing this field and renaming the file to {}.",
|
|
||||||
new_name,
|
|
||||||
path.display()
|
|
||||||
);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
instance.name = Some(name.to_owned());
|
|
||||||
|
|
||||||
let id = instance.id.take().map(RojoRef::new);
|
let id = instance.id.take().map(RojoRef::new);
|
||||||
let schema = instance.schema.take();
|
let schema = instance.schema.take();
|
||||||
|
|
||||||
@@ -62,7 +56,8 @@ pub fn snapshot_json_model(
|
|||||||
.relevant_paths(vec![vfs.canonicalize(path)?])
|
.relevant_paths(vec![vfs.canonicalize(path)?])
|
||||||
.context(context)
|
.context(context)
|
||||||
.specified_id(id)
|
.specified_id(id)
|
||||||
.schema(schema);
|
.schema(schema)
|
||||||
|
.specified_name(specified_name);
|
||||||
|
|
||||||
Ok(Some(snapshot))
|
Ok(Some(snapshot))
|
||||||
}
|
}
|
||||||
@@ -81,6 +76,7 @@ pub fn syncback_json_model<'sync>(
|
|||||||
// schemas will ever exist in one project for it to matter, but it
|
// schemas will ever exist in one project for it to matter, but it
|
||||||
// could have a performance cost.
|
// could have a performance cost.
|
||||||
model.schema = old_inst.metadata().schema.clone();
|
model.schema = old_inst.metadata().schema.clone();
|
||||||
|
model.name = old_inst.metadata().specified_name.clone();
|
||||||
}
|
}
|
||||||
|
|
||||||
Ok(SyncbackReturn {
|
Ok(SyncbackReturn {
|
||||||
|
|||||||
@@ -158,8 +158,16 @@ pub fn syncback_lua<'sync>(
|
|||||||
|
|
||||||
if !meta.is_empty() {
|
if !meta.is_empty() {
|
||||||
let parent_location = snapshot.path.parent_err()?;
|
let parent_location = snapshot.path.parent_err()?;
|
||||||
|
let instance_name = &snapshot.new_inst().name;
|
||||||
|
let slugified;
|
||||||
|
let meta_name = if crate::syncback::validate_file_name(instance_name).is_err() {
|
||||||
|
slugified = crate::syncback::slugify_name(instance_name);
|
||||||
|
&slugified
|
||||||
|
} else {
|
||||||
|
instance_name
|
||||||
|
};
|
||||||
fs_snapshot.add_file(
|
fs_snapshot.add_file(
|
||||||
parent_location.join(format!("{}.meta.json", new_inst.name)),
|
parent_location.join(format!("{}.meta.json", meta_name)),
|
||||||
serde_json::to_vec_pretty(&meta).context("cannot serialize metadata")?,
|
serde_json::to_vec_pretty(&meta).context("cannot serialize metadata")?,
|
||||||
);
|
);
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -10,7 +10,10 @@ use rbx_dom_weak::{
|
|||||||
use serde::{Deserialize, Serialize};
|
use serde::{Deserialize, Serialize};
|
||||||
|
|
||||||
use crate::{
|
use crate::{
|
||||||
json, resolution::UnresolvedValue, snapshot::InstanceSnapshot, syncback::SyncbackSnapshot,
|
json,
|
||||||
|
resolution::UnresolvedValue,
|
||||||
|
snapshot::InstanceSnapshot,
|
||||||
|
syncback::{validate_file_name, SyncbackSnapshot},
|
||||||
RojoRef,
|
RojoRef,
|
||||||
};
|
};
|
||||||
|
|
||||||
@@ -36,6 +39,9 @@ pub struct AdjacentMetadata {
|
|||||||
#[serde(default, skip_serializing_if = "IndexMap::is_empty")]
|
#[serde(default, skip_serializing_if = "IndexMap::is_empty")]
|
||||||
pub attributes: IndexMap<String, UnresolvedValue>,
|
pub attributes: IndexMap<String, UnresolvedValue>,
|
||||||
|
|
||||||
|
#[serde(skip_serializing_if = "Option::is_none")]
|
||||||
|
pub name: Option<String>,
|
||||||
|
|
||||||
#[serde(skip)]
|
#[serde(skip)]
|
||||||
pub path: PathBuf,
|
pub path: PathBuf,
|
||||||
}
|
}
|
||||||
@@ -144,6 +150,24 @@ impl AdjacentMetadata {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
let name = snapshot
|
||||||
|
.old_inst()
|
||||||
|
.and_then(|inst| inst.metadata().specified_name.clone())
|
||||||
|
.or_else(|| {
|
||||||
|
// If this is a new instance and its name is invalid for the filesystem,
|
||||||
|
// we need to specify the name in meta.json so it can be preserved
|
||||||
|
if snapshot.old_inst().is_none() {
|
||||||
|
let instance_name = &snapshot.new_inst().name;
|
||||||
|
if validate_file_name(instance_name).is_err() {
|
||||||
|
Some(instance_name.clone())
|
||||||
|
} else {
|
||||||
|
None
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
None
|
||||||
|
}
|
||||||
|
});
|
||||||
|
|
||||||
Ok(Some(Self {
|
Ok(Some(Self {
|
||||||
ignore_unknown_instances: if ignore_unknown_instances {
|
ignore_unknown_instances: if ignore_unknown_instances {
|
||||||
Some(true)
|
Some(true)
|
||||||
@@ -155,6 +179,7 @@ impl AdjacentMetadata {
|
|||||||
path,
|
path,
|
||||||
id: None,
|
id: None,
|
||||||
schema,
|
schema,
|
||||||
|
name,
|
||||||
}))
|
}))
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -213,11 +238,26 @@ impl AdjacentMetadata {
|
|||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
|
|
||||||
|
fn apply_name(&mut self, snapshot: &mut InstanceSnapshot) -> anyhow::Result<()> {
|
||||||
|
if self.name.is_some() && snapshot.metadata.specified_name.is_some() {
|
||||||
|
anyhow::bail!(
|
||||||
|
"cannot specify a name using {} (instance has a name from somewhere else)",
|
||||||
|
self.path.display()
|
||||||
|
);
|
||||||
|
}
|
||||||
|
if let Some(name) = &self.name {
|
||||||
|
snapshot.name = name.clone().into();
|
||||||
|
}
|
||||||
|
snapshot.metadata.specified_name = self.name.take();
|
||||||
|
Ok(())
|
||||||
|
}
|
||||||
|
|
||||||
pub fn apply_all(&mut self, snapshot: &mut InstanceSnapshot) -> anyhow::Result<()> {
|
pub fn apply_all(&mut self, snapshot: &mut InstanceSnapshot) -> anyhow::Result<()> {
|
||||||
self.apply_ignore_unknown_instances(snapshot);
|
self.apply_ignore_unknown_instances(snapshot);
|
||||||
self.apply_properties(snapshot)?;
|
self.apply_properties(snapshot)?;
|
||||||
self.apply_id(snapshot)?;
|
self.apply_id(snapshot)?;
|
||||||
self.apply_schema(snapshot)?;
|
self.apply_schema(snapshot)?;
|
||||||
|
self.apply_name(snapshot)?;
|
||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -226,11 +266,13 @@ impl AdjacentMetadata {
|
|||||||
///
|
///
|
||||||
/// - The number of properties and attributes is 0
|
/// - The number of properties and attributes is 0
|
||||||
/// - `ignore_unknown_instances` is None
|
/// - `ignore_unknown_instances` is None
|
||||||
|
/// - `name` is None
|
||||||
#[inline]
|
#[inline]
|
||||||
pub fn is_empty(&self) -> bool {
|
pub fn is_empty(&self) -> bool {
|
||||||
self.attributes.is_empty()
|
self.attributes.is_empty()
|
||||||
&& self.properties.is_empty()
|
&& self.properties.is_empty()
|
||||||
&& self.ignore_unknown_instances.is_none()
|
&& self.ignore_unknown_instances.is_none()
|
||||||
|
&& self.name.is_none()
|
||||||
}
|
}
|
||||||
|
|
||||||
// TODO: Add method to allow selectively applying parts of metadata and
|
// TODO: Add method to allow selectively applying parts of metadata and
|
||||||
@@ -262,6 +304,9 @@ pub struct DirectoryMetadata {
|
|||||||
#[serde(skip_serializing_if = "Option::is_none")]
|
#[serde(skip_serializing_if = "Option::is_none")]
|
||||||
pub class_name: Option<Ustr>,
|
pub class_name: Option<Ustr>,
|
||||||
|
|
||||||
|
#[serde(skip_serializing_if = "Option::is_none")]
|
||||||
|
pub name: Option<String>,
|
||||||
|
|
||||||
#[serde(skip)]
|
#[serde(skip)]
|
||||||
pub path: PathBuf,
|
pub path: PathBuf,
|
||||||
}
|
}
|
||||||
@@ -372,6 +417,24 @@ impl DirectoryMetadata {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
let name = snapshot
|
||||||
|
.old_inst()
|
||||||
|
.and_then(|inst| inst.metadata().specified_name.clone())
|
||||||
|
.or_else(|| {
|
||||||
|
// If this is a new instance and its name is invalid for the filesystem,
|
||||||
|
// we need to specify the name in meta.json so it can be preserved
|
||||||
|
if snapshot.old_inst().is_none() {
|
||||||
|
let instance_name = &snapshot.new_inst().name;
|
||||||
|
if validate_file_name(instance_name).is_err() {
|
||||||
|
Some(instance_name.clone())
|
||||||
|
} else {
|
||||||
|
None
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
None
|
||||||
|
}
|
||||||
|
});
|
||||||
|
|
||||||
Ok(Some(Self {
|
Ok(Some(Self {
|
||||||
ignore_unknown_instances: if ignore_unknown_instances {
|
ignore_unknown_instances: if ignore_unknown_instances {
|
||||||
Some(true)
|
Some(true)
|
||||||
@@ -384,6 +447,7 @@ impl DirectoryMetadata {
|
|||||||
path,
|
path,
|
||||||
id: None,
|
id: None,
|
||||||
schema,
|
schema,
|
||||||
|
name,
|
||||||
}))
|
}))
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -393,6 +457,7 @@ impl DirectoryMetadata {
|
|||||||
self.apply_properties(snapshot)?;
|
self.apply_properties(snapshot)?;
|
||||||
self.apply_id(snapshot)?;
|
self.apply_id(snapshot)?;
|
||||||
self.apply_schema(snapshot)?;
|
self.apply_schema(snapshot)?;
|
||||||
|
self.apply_name(snapshot)?;
|
||||||
|
|
||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
@@ -464,17 +529,33 @@ impl DirectoryMetadata {
|
|||||||
snapshot.metadata.schema = self.schema.take();
|
snapshot.metadata.schema = self.schema.take();
|
||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
|
|
||||||
|
fn apply_name(&mut self, snapshot: &mut InstanceSnapshot) -> anyhow::Result<()> {
|
||||||
|
if self.name.is_some() && snapshot.metadata.specified_name.is_some() {
|
||||||
|
anyhow::bail!(
|
||||||
|
"cannot specify a name using {} (instance has a name from somewhere else)",
|
||||||
|
self.path.display()
|
||||||
|
);
|
||||||
|
}
|
||||||
|
if let Some(name) = &self.name {
|
||||||
|
snapshot.name = name.clone().into();
|
||||||
|
}
|
||||||
|
snapshot.metadata.specified_name = self.name.take();
|
||||||
|
Ok(())
|
||||||
|
}
|
||||||
/// Returns whether the metadata is 'empty', meaning it doesn't have anything
|
/// Returns whether the metadata is 'empty', meaning it doesn't have anything
|
||||||
/// worth persisting in it. Specifically:
|
/// worth persisting in it. Specifically:
|
||||||
///
|
///
|
||||||
/// - The number of properties and attributes is 0
|
/// - The number of properties and attributes is 0
|
||||||
/// - `ignore_unknown_instances` is None
|
/// - `ignore_unknown_instances` is None
|
||||||
/// - `class_name` is either None or not Some("Folder")
|
/// - `class_name` is either None or not Some("Folder")
|
||||||
|
/// - `name` is None
|
||||||
#[inline]
|
#[inline]
|
||||||
pub fn is_empty(&self) -> bool {
|
pub fn is_empty(&self) -> bool {
|
||||||
self.attributes.is_empty()
|
self.attributes.is_empty()
|
||||||
&& self.properties.is_empty()
|
&& self.properties.is_empty()
|
||||||
&& self.ignore_unknown_instances.is_none()
|
&& self.ignore_unknown_instances.is_none()
|
||||||
|
&& self.name.is_none()
|
||||||
&& if let Some(class) = &self.class_name {
|
&& if let Some(class) = &self.class_name {
|
||||||
class == "Folder"
|
class == "Folder"
|
||||||
} else {
|
} else {
|
||||||
|
|||||||
@@ -8,11 +8,11 @@ use rbx_dom_weak::Instance;
|
|||||||
|
|
||||||
use crate::{snapshot::InstanceWithMeta, snapshot_middleware::Middleware};
|
use crate::{snapshot::InstanceWithMeta, snapshot_middleware::Middleware};
|
||||||
|
|
||||||
pub fn name_for_inst<'old>(
|
pub fn name_for_inst<'a>(
|
||||||
middleware: Middleware,
|
middleware: Middleware,
|
||||||
new_inst: &Instance,
|
new_inst: &'a Instance,
|
||||||
old_inst: Option<InstanceWithMeta<'old>>,
|
old_inst: Option<InstanceWithMeta<'a>>,
|
||||||
) -> anyhow::Result<Cow<'old, str>> {
|
) -> anyhow::Result<Cow<'a, str>> {
|
||||||
if let Some(old_inst) = old_inst {
|
if let Some(old_inst) = old_inst {
|
||||||
if let Some(source) = old_inst.metadata().relevant_paths.first() {
|
if let Some(source) = old_inst.metadata().relevant_paths.first() {
|
||||||
source
|
source
|
||||||
@@ -35,14 +35,24 @@ pub fn name_for_inst<'old>(
|
|||||||
| Middleware::CsvDir
|
| Middleware::CsvDir
|
||||||
| Middleware::ServerScriptDir
|
| Middleware::ServerScriptDir
|
||||||
| Middleware::ClientScriptDir
|
| Middleware::ClientScriptDir
|
||||||
| Middleware::ModuleScriptDir => Cow::Owned(new_inst.name.clone()),
|
| Middleware::ModuleScriptDir => {
|
||||||
|
if validate_file_name(&new_inst.name).is_err() {
|
||||||
|
Cow::Owned(slugify_name(&new_inst.name))
|
||||||
|
} else {
|
||||||
|
Cow::Borrowed(&new_inst.name)
|
||||||
|
}
|
||||||
|
}
|
||||||
_ => {
|
_ => {
|
||||||
let extension = extension_for_middleware(middleware);
|
let extension = extension_for_middleware(middleware);
|
||||||
let name = &new_inst.name;
|
let slugified;
|
||||||
validate_file_name(name).with_context(|| {
|
let final_name = if validate_file_name(&new_inst.name).is_err() {
|
||||||
format!("name '{name}' is not legal to write to the file system")
|
slugified = slugify_name(&new_inst.name);
|
||||||
})?;
|
&slugified
|
||||||
Cow::Owned(format!("{name}.{extension}"))
|
} else {
|
||||||
|
&new_inst.name
|
||||||
|
};
|
||||||
|
|
||||||
|
Cow::Owned(format!("{final_name}.{extension}"))
|
||||||
}
|
}
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
@@ -94,6 +104,39 @@ const INVALID_WINDOWS_NAMES: [&str; 22] = [
|
|||||||
/// in a file's name.
|
/// in a file's name.
|
||||||
const FORBIDDEN_CHARS: [char; 9] = ['<', '>', ':', '"', '/', '|', '?', '*', '\\'];
|
const FORBIDDEN_CHARS: [char; 9] = ['<', '>', ':', '"', '/', '|', '?', '*', '\\'];
|
||||||
|
|
||||||
|
/// Slugifies a name by replacing forbidden characters with underscores
|
||||||
|
/// and ensuring the result is a valid file name
|
||||||
|
pub fn slugify_name(name: &str) -> String {
|
||||||
|
let mut result = String::with_capacity(name.len());
|
||||||
|
|
||||||
|
for ch in name.chars() {
|
||||||
|
if FORBIDDEN_CHARS.contains(&ch) {
|
||||||
|
result.push('_');
|
||||||
|
} else {
|
||||||
|
result.push(ch);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Handle Windows reserved names by appending an underscore
|
||||||
|
let result_lower = result.to_lowercase();
|
||||||
|
for forbidden in INVALID_WINDOWS_NAMES {
|
||||||
|
if result_lower == forbidden.to_lowercase() {
|
||||||
|
result.push('_');
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
while result.ends_with(' ') || result.ends_with('.') {
|
||||||
|
result.pop();
|
||||||
|
}
|
||||||
|
|
||||||
|
if result.is_empty() || result.chars().all(|c| c == '_') {
|
||||||
|
result = "instance".to_string();
|
||||||
|
}
|
||||||
|
|
||||||
|
result
|
||||||
|
}
|
||||||
|
|
||||||
/// Validates a provided file name to ensure it's allowed on the file system. An
|
/// Validates a provided file name to ensure it's allowed on the file system. An
|
||||||
/// error is returned if the name isn't allowed, indicating why.
|
/// error is returned if the name isn't allowed, indicating why.
|
||||||
/// This takes into account rules for Windows, MacOS, and Linux.
|
/// This takes into account rules for Windows, MacOS, and Linux.
|
||||||
|
|||||||
@@ -28,7 +28,7 @@ use crate::{
|
|||||||
Project,
|
Project,
|
||||||
};
|
};
|
||||||
|
|
||||||
pub use file_names::{extension_for_middleware, name_for_inst, validate_file_name};
|
pub use file_names::{extension_for_middleware, name_for_inst, slugify_name, validate_file_name};
|
||||||
pub use fs_snapshot::FsSnapshot;
|
pub use fs_snapshot::FsSnapshot;
|
||||||
pub use hash::*;
|
pub use hash::*;
|
||||||
pub use property_filter::{filter_properties, filter_properties_preallocated};
|
pub use property_filter::{filter_properties, filter_properties_preallocated};
|
||||||
@@ -301,7 +301,6 @@ pub fn get_best_middleware(snapshot: &SyncbackSnapshot) -> Middleware {
|
|||||||
static JSON_MODEL_CLASSES: OnceLock<HashSet<&str>> = OnceLock::new();
|
static JSON_MODEL_CLASSES: OnceLock<HashSet<&str>> = OnceLock::new();
|
||||||
let json_model_classes = JSON_MODEL_CLASSES.get_or_init(|| {
|
let json_model_classes = JSON_MODEL_CLASSES.get_or_init(|| {
|
||||||
[
|
[
|
||||||
"Actor",
|
|
||||||
"Sound",
|
"Sound",
|
||||||
"SoundGroup",
|
"SoundGroup",
|
||||||
"Sky",
|
"Sky",
|
||||||
@@ -319,11 +318,6 @@ pub fn get_best_middleware(snapshot: &SyncbackSnapshot) -> Middleware {
|
|||||||
"ChatInputBarConfiguration",
|
"ChatInputBarConfiguration",
|
||||||
"BubbleChatConfiguration",
|
"BubbleChatConfiguration",
|
||||||
"ChannelTabsConfiguration",
|
"ChannelTabsConfiguration",
|
||||||
"RemoteEvent",
|
|
||||||
"UnreliableRemoteEvent",
|
|
||||||
"RemoteFunction",
|
|
||||||
"BindableEvent",
|
|
||||||
"BindableFunction",
|
|
||||||
]
|
]
|
||||||
.into()
|
.into()
|
||||||
});
|
});
|
||||||
|
|||||||
@@ -13,6 +13,7 @@ use rbx_dom_weak::{
|
|||||||
};
|
};
|
||||||
|
|
||||||
use crate::{
|
use crate::{
|
||||||
|
json,
|
||||||
serve_session::ServeSession,
|
serve_session::ServeSession,
|
||||||
snapshot::{InstanceWithMeta, PatchSet, PatchUpdate},
|
snapshot::{InstanceWithMeta, PatchSet, PatchUpdate},
|
||||||
web::{
|
web::{
|
||||||
@@ -21,10 +22,11 @@ use crate::{
|
|||||||
ServerInfoResponse, SocketPacket, SocketPacketBody, SocketPacketType, SubscribeMessage,
|
ServerInfoResponse, SocketPacket, SocketPacketBody, SocketPacketType, SubscribeMessage,
|
||||||
WriteRequest, WriteResponse, PROTOCOL_VERSION, SERVER_VERSION,
|
WriteRequest, WriteResponse, PROTOCOL_VERSION, SERVER_VERSION,
|
||||||
},
|
},
|
||||||
util::{deserialize_msgpack, msgpack, msgpack_ok, serialize_msgpack},
|
util::{json, json_ok},
|
||||||
},
|
},
|
||||||
web_api::{
|
web_api::{
|
||||||
InstanceUpdate, RefPatchRequest, RefPatchResponse, SerializeRequest, SerializeResponse,
|
BufferEncode, InstanceUpdate, RefPatchRequest, RefPatchResponse, SerializeRequest,
|
||||||
|
SerializeResponse,
|
||||||
},
|
},
|
||||||
};
|
};
|
||||||
|
|
||||||
@@ -40,7 +42,7 @@ pub async fn call(serve_session: Arc<ServeSession>, mut request: Request<Body>)
|
|||||||
if is_upgrade_request(&request) {
|
if is_upgrade_request(&request) {
|
||||||
service.handle_api_socket(&mut request).await
|
service.handle_api_socket(&mut request).await
|
||||||
} else {
|
} else {
|
||||||
msgpack(
|
json(
|
||||||
ErrorResponse::bad_request(
|
ErrorResponse::bad_request(
|
||||||
"/api/socket must be called as a websocket upgrade request",
|
"/api/socket must be called as a websocket upgrade request",
|
||||||
),
|
),
|
||||||
@@ -56,7 +58,7 @@ pub async fn call(serve_session: Arc<ServeSession>, mut request: Request<Body>)
|
|||||||
}
|
}
|
||||||
(&Method::POST, "/api/write") => service.handle_api_write(request).await,
|
(&Method::POST, "/api/write") => service.handle_api_write(request).await,
|
||||||
|
|
||||||
(_method, path) => msgpack(
|
(_method, path) => json(
|
||||||
ErrorResponse::not_found(format!("Route not found: {}", path)),
|
ErrorResponse::not_found(format!("Route not found: {}", path)),
|
||||||
StatusCode::NOT_FOUND,
|
StatusCode::NOT_FOUND,
|
||||||
),
|
),
|
||||||
@@ -77,7 +79,7 @@ impl ApiService {
|
|||||||
let tree = self.serve_session.tree();
|
let tree = self.serve_session.tree();
|
||||||
let root_instance_id = tree.get_root_id();
|
let root_instance_id = tree.get_root_id();
|
||||||
|
|
||||||
msgpack_ok(&ServerInfoResponse {
|
json_ok(&ServerInfoResponse {
|
||||||
server_version: SERVER_VERSION.to_owned(),
|
server_version: SERVER_VERSION.to_owned(),
|
||||||
protocol_version: PROTOCOL_VERSION,
|
protocol_version: PROTOCOL_VERSION,
|
||||||
session_id: self.serve_session.session_id(),
|
session_id: self.serve_session.session_id(),
|
||||||
@@ -96,7 +98,7 @@ impl ApiService {
|
|||||||
let input_cursor: u32 = match argument.parse() {
|
let input_cursor: u32 = match argument.parse() {
|
||||||
Ok(v) => v,
|
Ok(v) => v,
|
||||||
Err(err) => {
|
Err(err) => {
|
||||||
return msgpack(
|
return json(
|
||||||
ErrorResponse::bad_request(format!("Malformed message cursor: {}", err)),
|
ErrorResponse::bad_request(format!("Malformed message cursor: {}", err)),
|
||||||
StatusCode::BAD_REQUEST,
|
StatusCode::BAD_REQUEST,
|
||||||
);
|
);
|
||||||
@@ -107,7 +109,7 @@ impl ApiService {
|
|||||||
let (response, websocket) = match upgrade(request, None) {
|
let (response, websocket) = match upgrade(request, None) {
|
||||||
Ok(result) => result,
|
Ok(result) => result,
|
||||||
Err(err) => {
|
Err(err) => {
|
||||||
return msgpack(
|
return json(
|
||||||
ErrorResponse::internal_error(format!("WebSocket upgrade failed: {}", err)),
|
ErrorResponse::internal_error(format!("WebSocket upgrade failed: {}", err)),
|
||||||
StatusCode::INTERNAL_SERVER_ERROR,
|
StatusCode::INTERNAL_SERVER_ERROR,
|
||||||
);
|
);
|
||||||
@@ -134,10 +136,10 @@ impl ApiService {
|
|||||||
|
|
||||||
let body = body::to_bytes(request.into_body()).await.unwrap();
|
let body = body::to_bytes(request.into_body()).await.unwrap();
|
||||||
|
|
||||||
let request: WriteRequest = match deserialize_msgpack(&body) {
|
let request: WriteRequest = match json::from_slice(&body) {
|
||||||
Ok(request) => request,
|
Ok(request) => request,
|
||||||
Err(err) => {
|
Err(err) => {
|
||||||
return msgpack(
|
return json(
|
||||||
ErrorResponse::bad_request(format!("Invalid body: {}", err)),
|
ErrorResponse::bad_request(format!("Invalid body: {}", err)),
|
||||||
StatusCode::BAD_REQUEST,
|
StatusCode::BAD_REQUEST,
|
||||||
);
|
);
|
||||||
@@ -145,7 +147,7 @@ impl ApiService {
|
|||||||
};
|
};
|
||||||
|
|
||||||
if request.session_id != session_id {
|
if request.session_id != session_id {
|
||||||
return msgpack(
|
return json(
|
||||||
ErrorResponse::bad_request("Wrong session ID"),
|
ErrorResponse::bad_request("Wrong session ID"),
|
||||||
StatusCode::BAD_REQUEST,
|
StatusCode::BAD_REQUEST,
|
||||||
);
|
);
|
||||||
@@ -171,7 +173,7 @@ impl ApiService {
|
|||||||
})
|
})
|
||||||
.unwrap();
|
.unwrap();
|
||||||
|
|
||||||
msgpack_ok(WriteResponse { session_id })
|
json_ok(WriteResponse { session_id })
|
||||||
}
|
}
|
||||||
|
|
||||||
async fn handle_api_read(&self, request: Request<Body>) -> Response<Body> {
|
async fn handle_api_read(&self, request: Request<Body>) -> Response<Body> {
|
||||||
@@ -181,7 +183,7 @@ impl ApiService {
|
|||||||
let requested_ids = match requested_ids {
|
let requested_ids = match requested_ids {
|
||||||
Ok(ids) => ids,
|
Ok(ids) => ids,
|
||||||
Err(_) => {
|
Err(_) => {
|
||||||
return msgpack(
|
return json(
|
||||||
ErrorResponse::bad_request("Malformed ID list"),
|
ErrorResponse::bad_request("Malformed ID list"),
|
||||||
StatusCode::BAD_REQUEST,
|
StatusCode::BAD_REQUEST,
|
||||||
);
|
);
|
||||||
@@ -205,7 +207,7 @@ impl ApiService {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
msgpack_ok(ReadResponse {
|
json_ok(ReadResponse {
|
||||||
session_id: self.serve_session.session_id(),
|
session_id: self.serve_session.session_id(),
|
||||||
message_cursor,
|
message_cursor,
|
||||||
instances,
|
instances,
|
||||||
@@ -223,10 +225,10 @@ impl ApiService {
|
|||||||
let session_id = self.serve_session.session_id();
|
let session_id = self.serve_session.session_id();
|
||||||
let body = body::to_bytes(request.into_body()).await.unwrap();
|
let body = body::to_bytes(request.into_body()).await.unwrap();
|
||||||
|
|
||||||
let request: SerializeRequest = match deserialize_msgpack(&body) {
|
let request: SerializeRequest = match json::from_slice(&body) {
|
||||||
Ok(request) => request,
|
Ok(request) => request,
|
||||||
Err(err) => {
|
Err(err) => {
|
||||||
return msgpack(
|
return json(
|
||||||
ErrorResponse::bad_request(format!("Invalid body: {}", err)),
|
ErrorResponse::bad_request(format!("Invalid body: {}", err)),
|
||||||
StatusCode::BAD_REQUEST,
|
StatusCode::BAD_REQUEST,
|
||||||
);
|
);
|
||||||
@@ -234,7 +236,7 @@ impl ApiService {
|
|||||||
};
|
};
|
||||||
|
|
||||||
if request.session_id != session_id {
|
if request.session_id != session_id {
|
||||||
return msgpack(
|
return json(
|
||||||
ErrorResponse::bad_request("Wrong session ID"),
|
ErrorResponse::bad_request("Wrong session ID"),
|
||||||
StatusCode::BAD_REQUEST,
|
StatusCode::BAD_REQUEST,
|
||||||
);
|
);
|
||||||
@@ -267,7 +269,7 @@ impl ApiService {
|
|||||||
|
|
||||||
response_dom.transfer_within(child_ref, object_value);
|
response_dom.transfer_within(child_ref, object_value);
|
||||||
} else {
|
} else {
|
||||||
msgpack(
|
json(
|
||||||
ErrorResponse::bad_request(format!("provided id {id} is not in the tree")),
|
ErrorResponse::bad_request(format!("provided id {id} is not in the tree")),
|
||||||
StatusCode::BAD_REQUEST,
|
StatusCode::BAD_REQUEST,
|
||||||
);
|
);
|
||||||
@@ -278,9 +280,9 @@ impl ApiService {
|
|||||||
let mut source = Vec::new();
|
let mut source = Vec::new();
|
||||||
rbx_binary::to_writer(&mut source, &response_dom, &[response_dom.root_ref()]).unwrap();
|
rbx_binary::to_writer(&mut source, &response_dom, &[response_dom.root_ref()]).unwrap();
|
||||||
|
|
||||||
msgpack_ok(SerializeResponse {
|
json_ok(SerializeResponse {
|
||||||
session_id: self.serve_session.session_id(),
|
session_id: self.serve_session.session_id(),
|
||||||
model_contents: source,
|
model_contents: BufferEncode::new(source),
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -292,10 +294,10 @@ impl ApiService {
|
|||||||
let session_id = self.serve_session.session_id();
|
let session_id = self.serve_session.session_id();
|
||||||
let body = body::to_bytes(request.into_body()).await.unwrap();
|
let body = body::to_bytes(request.into_body()).await.unwrap();
|
||||||
|
|
||||||
let request: RefPatchRequest = match deserialize_msgpack(&body) {
|
let request: RefPatchRequest = match json::from_slice(&body) {
|
||||||
Ok(request) => request,
|
Ok(request) => request,
|
||||||
Err(err) => {
|
Err(err) => {
|
||||||
return msgpack(
|
return json(
|
||||||
ErrorResponse::bad_request(format!("Invalid body: {}", err)),
|
ErrorResponse::bad_request(format!("Invalid body: {}", err)),
|
||||||
StatusCode::BAD_REQUEST,
|
StatusCode::BAD_REQUEST,
|
||||||
);
|
);
|
||||||
@@ -303,7 +305,7 @@ impl ApiService {
|
|||||||
};
|
};
|
||||||
|
|
||||||
if request.session_id != session_id {
|
if request.session_id != session_id {
|
||||||
return msgpack(
|
return json(
|
||||||
ErrorResponse::bad_request("Wrong session ID"),
|
ErrorResponse::bad_request("Wrong session ID"),
|
||||||
StatusCode::BAD_REQUEST,
|
StatusCode::BAD_REQUEST,
|
||||||
);
|
);
|
||||||
@@ -336,7 +338,7 @@ impl ApiService {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
msgpack_ok(RefPatchResponse {
|
json_ok(RefPatchResponse {
|
||||||
session_id: self.serve_session.session_id(),
|
session_id: self.serve_session.session_id(),
|
||||||
patch: SubscribeMessage {
|
patch: SubscribeMessage {
|
||||||
added: HashMap::new(),
|
added: HashMap::new(),
|
||||||
@@ -352,7 +354,7 @@ impl ApiService {
|
|||||||
let requested_id = match Ref::from_str(argument) {
|
let requested_id = match Ref::from_str(argument) {
|
||||||
Ok(id) => id,
|
Ok(id) => id,
|
||||||
Err(_) => {
|
Err(_) => {
|
||||||
return msgpack(
|
return json(
|
||||||
ErrorResponse::bad_request("Invalid instance ID"),
|
ErrorResponse::bad_request("Invalid instance ID"),
|
||||||
StatusCode::BAD_REQUEST,
|
StatusCode::BAD_REQUEST,
|
||||||
);
|
);
|
||||||
@@ -364,7 +366,7 @@ impl ApiService {
|
|||||||
let instance = match tree.get_instance(requested_id) {
|
let instance = match tree.get_instance(requested_id) {
|
||||||
Some(instance) => instance,
|
Some(instance) => instance,
|
||||||
None => {
|
None => {
|
||||||
return msgpack(
|
return json(
|
||||||
ErrorResponse::bad_request("Instance not found"),
|
ErrorResponse::bad_request("Instance not found"),
|
||||||
StatusCode::NOT_FOUND,
|
StatusCode::NOT_FOUND,
|
||||||
);
|
);
|
||||||
@@ -374,7 +376,7 @@ impl ApiService {
|
|||||||
let script_path = match pick_script_path(instance) {
|
let script_path = match pick_script_path(instance) {
|
||||||
Some(path) => path,
|
Some(path) => path,
|
||||||
None => {
|
None => {
|
||||||
return msgpack(
|
return json(
|
||||||
ErrorResponse::bad_request(
|
ErrorResponse::bad_request(
|
||||||
"No appropriate file could be found to open this script",
|
"No appropriate file could be found to open this script",
|
||||||
),
|
),
|
||||||
@@ -387,7 +389,7 @@ impl ApiService {
|
|||||||
Ok(()) => {}
|
Ok(()) => {}
|
||||||
Err(error) => match error {
|
Err(error) => match error {
|
||||||
OpenError::Io(io_error) => {
|
OpenError::Io(io_error) => {
|
||||||
return msgpack(
|
return json(
|
||||||
ErrorResponse::internal_error(format!(
|
ErrorResponse::internal_error(format!(
|
||||||
"Attempting to open {} failed because of the following io error: {}",
|
"Attempting to open {} failed because of the following io error: {}",
|
||||||
script_path.display(),
|
script_path.display(),
|
||||||
@@ -401,7 +403,7 @@ impl ApiService {
|
|||||||
status,
|
status,
|
||||||
stderr,
|
stderr,
|
||||||
} => {
|
} => {
|
||||||
return msgpack(
|
return json(
|
||||||
ErrorResponse::internal_error(format!(
|
ErrorResponse::internal_error(format!(
|
||||||
r#"The command '{}' to open '{}' failed with the error code '{}'.
|
r#"The command '{}' to open '{}' failed with the error code '{}'.
|
||||||
Error logs:
|
Error logs:
|
||||||
@@ -417,7 +419,7 @@ impl ApiService {
|
|||||||
},
|
},
|
||||||
};
|
};
|
||||||
|
|
||||||
msgpack_ok(OpenResponse {
|
json_ok(OpenResponse {
|
||||||
session_id: self.serve_session.session_id(),
|
session_id: self.serve_session.session_id(),
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
@@ -481,7 +483,7 @@ async fn handle_websocket_subscription(
|
|||||||
match result {
|
match result {
|
||||||
Ok((new_cursor, messages)) => {
|
Ok((new_cursor, messages)) => {
|
||||||
if !messages.is_empty() {
|
if !messages.is_empty() {
|
||||||
let msgpack_message = {
|
let json_message = {
|
||||||
let tree = tree_handle.lock().unwrap();
|
let tree = tree_handle.lock().unwrap();
|
||||||
let api_messages = messages
|
let api_messages = messages
|
||||||
.into_iter()
|
.into_iter()
|
||||||
@@ -497,12 +499,12 @@ async fn handle_websocket_subscription(
|
|||||||
}),
|
}),
|
||||||
};
|
};
|
||||||
|
|
||||||
serialize_msgpack(response)?
|
serde_json::to_string(&response)?
|
||||||
};
|
};
|
||||||
|
|
||||||
log::debug!("Sending batch of messages over WebSocket subscription");
|
log::debug!("Sending batch of messages over WebSocket subscription");
|
||||||
|
|
||||||
if websocket.send(Message::Binary(msgpack_message)).await.is_err() {
|
if websocket.send(Message::Text(json_message)).await.is_err() {
|
||||||
// Client disconnected
|
// Client disconnected
|
||||||
log::debug!("WebSocket subscription closed by client");
|
log::debug!("WebSocket subscription closed by client");
|
||||||
break;
|
break;
|
||||||
|
|||||||
@@ -249,8 +249,31 @@ pub struct SerializeRequest {
|
|||||||
#[serde(rename_all = "camelCase")]
|
#[serde(rename_all = "camelCase")]
|
||||||
pub struct SerializeResponse {
|
pub struct SerializeResponse {
|
||||||
pub session_id: SessionId,
|
pub session_id: SessionId,
|
||||||
#[serde(with = "serde_bytes")]
|
pub model_contents: BufferEncode,
|
||||||
pub model_contents: Vec<u8>,
|
}
|
||||||
|
|
||||||
|
/// Using this struct we can force Roblox to JSONDecode this as a buffer.
|
||||||
|
/// This is what Roblox's serde APIs use, so it saves a step in the plugin.
|
||||||
|
#[derive(Debug, Serialize, Deserialize)]
|
||||||
|
pub struct BufferEncode {
|
||||||
|
m: (),
|
||||||
|
t: Cow<'static, str>,
|
||||||
|
base64: String,
|
||||||
|
}
|
||||||
|
|
||||||
|
impl BufferEncode {
|
||||||
|
pub fn new(content: Vec<u8>) -> Self {
|
||||||
|
let base64 = data_encoding::BASE64.encode(&content);
|
||||||
|
Self {
|
||||||
|
m: (),
|
||||||
|
t: Cow::Borrowed("buffer"),
|
||||||
|
base64,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
pub fn model(&self) -> &str {
|
||||||
|
&self.base64
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
#[derive(Debug, Serialize, Deserialize)]
|
#[derive(Debug, Serialize, Deserialize)]
|
||||||
|
|||||||
@@ -1,48 +1,8 @@
|
|||||||
use hyper::{header::CONTENT_TYPE, Body, Response, StatusCode};
|
use hyper::{header::CONTENT_TYPE, Body, Response, StatusCode};
|
||||||
use serde::{Deserialize, Serialize};
|
use serde::Serialize;
|
||||||
|
|
||||||
pub fn msgpack_ok<T: Serialize>(value: T) -> Response<Body> {
|
pub fn json_ok<T: Serialize>(value: T) -> Response<Body> {
|
||||||
msgpack(value, StatusCode::OK)
|
json(value, StatusCode::OK)
|
||||||
}
|
|
||||||
|
|
||||||
pub fn msgpack<T: Serialize>(value: T, code: StatusCode) -> Response<Body> {
|
|
||||||
let mut serialized = Vec::new();
|
|
||||||
let mut serializer = rmp_serde::Serializer::new(&mut serialized)
|
|
||||||
.with_human_readable()
|
|
||||||
.with_struct_map();
|
|
||||||
|
|
||||||
if let Err(err) = value.serialize(&mut serializer) {
|
|
||||||
return Response::builder()
|
|
||||||
.status(StatusCode::INTERNAL_SERVER_ERROR)
|
|
||||||
.header(CONTENT_TYPE, "text/plain")
|
|
||||||
.body(Body::from(err.to_string()))
|
|
||||||
.unwrap();
|
|
||||||
};
|
|
||||||
|
|
||||||
Response::builder()
|
|
||||||
.status(code)
|
|
||||||
.header(CONTENT_TYPE, "application/msgpack")
|
|
||||||
.body(Body::from(serialized))
|
|
||||||
.unwrap()
|
|
||||||
}
|
|
||||||
|
|
||||||
pub fn serialize_msgpack<T: Serialize>(value: T) -> anyhow::Result<Vec<u8>> {
|
|
||||||
let mut serialized = Vec::new();
|
|
||||||
let mut serializer = rmp_serde::Serializer::new(&mut serialized)
|
|
||||||
.with_human_readable()
|
|
||||||
.with_struct_map();
|
|
||||||
|
|
||||||
value.serialize(&mut serializer)?;
|
|
||||||
|
|
||||||
Ok(serialized)
|
|
||||||
}
|
|
||||||
|
|
||||||
pub fn deserialize_msgpack<'a, T: Deserialize<'a>>(
|
|
||||||
input: &'a [u8],
|
|
||||||
) -> Result<T, rmp_serde::decode::Error> {
|
|
||||||
let mut deserializer = rmp_serde::Deserializer::new(input).with_human_readable();
|
|
||||||
|
|
||||||
T::deserialize(&mut deserializer)
|
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn json<T: Serialize>(value: T, code: StatusCode) -> Response<Body> {
|
pub fn json<T: Serialize>(value: T, code: StatusCode) -> Response<Body> {
|
||||||
|
|||||||
@@ -1,14 +0,0 @@
|
|||||||
{
|
|
||||||
"name": "default",
|
|
||||||
"tree": {
|
|
||||||
"$className": "DataModel",
|
|
||||||
"ReplicatedStorage": {
|
|
||||||
"Project": {
|
|
||||||
"$path": "project/src",
|
|
||||||
"Module": {
|
|
||||||
"$path": "module"
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
@@ -1 +0,0 @@
|
|||||||
return nil
|
|
||||||
@@ -1,14 +0,0 @@
|
|||||||
{
|
|
||||||
"name": "default",
|
|
||||||
"tree": {
|
|
||||||
"$className": "DataModel",
|
|
||||||
"ReplicatedStorage": {
|
|
||||||
"Project": {
|
|
||||||
"$path": "src/",
|
|
||||||
"Module": {
|
|
||||||
"$path": "../module"
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
@@ -1 +0,0 @@
|
|||||||
return nil
|
|
||||||
@@ -10,7 +10,6 @@ use std::{
|
|||||||
use hyper_tungstenite::tungstenite::{connect, Message};
|
use hyper_tungstenite::tungstenite::{connect, Message};
|
||||||
use rbx_dom_weak::types::Ref;
|
use rbx_dom_weak::types::Ref;
|
||||||
|
|
||||||
use serde::{Deserialize, Serialize};
|
|
||||||
use tempfile::{tempdir, TempDir};
|
use tempfile::{tempdir, TempDir};
|
||||||
|
|
||||||
use librojo::{
|
use librojo::{
|
||||||
@@ -162,16 +161,22 @@ impl TestServeSession {
|
|||||||
|
|
||||||
pub fn get_api_rojo(&self) -> Result<ServerInfoResponse, reqwest::Error> {
|
pub fn get_api_rojo(&self) -> Result<ServerInfoResponse, reqwest::Error> {
|
||||||
let url = format!("http://localhost:{}/api/rojo", self.port);
|
let url = format!("http://localhost:{}/api/rojo", self.port);
|
||||||
let body = reqwest::blocking::get(url)?.bytes()?;
|
let body = reqwest::blocking::get(url)?.text()?;
|
||||||
|
|
||||||
Ok(deserialize_msgpack(&body).expect("Server returned malformed response"))
|
let value = jsonc_parser::parse_to_serde_value(&body, &Default::default())
|
||||||
|
.expect("Failed to parse JSON")
|
||||||
|
.expect("No JSON value");
|
||||||
|
Ok(serde_json::from_value(value).expect("Server returned malformed response"))
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn get_api_read(&self, id: Ref) -> Result<ReadResponse<'_>, reqwest::Error> {
|
pub fn get_api_read(&self, id: Ref) -> Result<ReadResponse<'_>, reqwest::Error> {
|
||||||
let url = format!("http://localhost:{}/api/read/{}", self.port, id);
|
let url = format!("http://localhost:{}/api/read/{}", self.port, id);
|
||||||
let body = reqwest::blocking::get(url)?.bytes()?;
|
let body = reqwest::blocking::get(url)?.text()?;
|
||||||
|
|
||||||
Ok(deserialize_msgpack(&body).expect("Server returned malformed response"))
|
let value = jsonc_parser::parse_to_serde_value(&body, &Default::default())
|
||||||
|
.expect("Failed to parse JSON")
|
||||||
|
.expect("No JSON value");
|
||||||
|
Ok(serde_json::from_value(value).expect("Server returned malformed response"))
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn get_api_socket_packet(
|
pub fn get_api_socket_packet(
|
||||||
@@ -193,8 +198,8 @@ impl TestServeSession {
|
|||||||
}
|
}
|
||||||
|
|
||||||
match socket.read() {
|
match socket.read() {
|
||||||
Ok(Message::Binary(binary)) => {
|
Ok(Message::Text(text)) => {
|
||||||
let packet: SocketPacket = deserialize_msgpack(&binary)?;
|
let packet: SocketPacket = serde_json::from_str(&text)?;
|
||||||
if packet.packet_type != packet_type {
|
if packet.packet_type != packet_type {
|
||||||
continue;
|
continue;
|
||||||
}
|
}
|
||||||
@@ -207,7 +212,7 @@ impl TestServeSession {
|
|||||||
return Err("WebSocket closed before receiving messages".into());
|
return Err("WebSocket closed before receiving messages".into());
|
||||||
}
|
}
|
||||||
Ok(_) => {
|
Ok(_) => {
|
||||||
// Ignore other message types (ping, pong, text)
|
// Ignore other message types (ping, pong, binary)
|
||||||
continue;
|
continue;
|
||||||
}
|
}
|
||||||
Err(hyper_tungstenite::tungstenite::Error::Io(e))
|
Err(hyper_tungstenite::tungstenite::Error::Io(e))
|
||||||
@@ -231,37 +236,15 @@ impl TestServeSession {
|
|||||||
) -> Result<SerializeResponse, reqwest::Error> {
|
) -> Result<SerializeResponse, reqwest::Error> {
|
||||||
let client = reqwest::blocking::Client::new();
|
let client = reqwest::blocking::Client::new();
|
||||||
let url = format!("http://localhost:{}/api/serialize", self.port);
|
let url = format!("http://localhost:{}/api/serialize", self.port);
|
||||||
let body = serialize_msgpack(&SerializeRequest {
|
let body = serde_json::to_string(&SerializeRequest {
|
||||||
session_id,
|
session_id,
|
||||||
ids: ids.to_vec(),
|
ids: ids.to_vec(),
|
||||||
})
|
});
|
||||||
.unwrap();
|
|
||||||
|
|
||||||
let body = client.post(url).body(body).send()?.bytes()?;
|
client.post(url).body((body).unwrap()).send()?.json()
|
||||||
|
|
||||||
Ok(deserialize_msgpack(&body).expect("Server returned malformed response"))
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
fn serialize_msgpack<T: Serialize>(value: T) -> Result<Vec<u8>, rmp_serde::encode::Error> {
|
|
||||||
let mut serialized = Vec::new();
|
|
||||||
let mut serializer = rmp_serde::Serializer::new(&mut serialized)
|
|
||||||
.with_human_readable()
|
|
||||||
.with_struct_map();
|
|
||||||
|
|
||||||
value.serialize(&mut serializer)?;
|
|
||||||
|
|
||||||
Ok(serialized)
|
|
||||||
}
|
|
||||||
|
|
||||||
fn deserialize_msgpack<'a, T: Deserialize<'a>>(
|
|
||||||
input: &'a [u8],
|
|
||||||
) -> Result<T, rmp_serde::decode::Error> {
|
|
||||||
let mut deserializer = rmp_serde::Deserializer::new(input).with_human_readable();
|
|
||||||
|
|
||||||
T::deserialize(&mut deserializer)
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Probably-okay way to generate random enough port numbers for running the
|
/// Probably-okay way to generate random enough port numbers for running the
|
||||||
/// Rojo live server.
|
/// Rojo live server.
|
||||||
///
|
///
|
||||||
@@ -279,7 +262,11 @@ fn get_port_number() -> usize {
|
|||||||
/// Since the provided structure intentionally includes unredacted referents,
|
/// Since the provided structure intentionally includes unredacted referents,
|
||||||
/// some post-processing is done to ensure they don't show up in the model.
|
/// some post-processing is done to ensure they don't show up in the model.
|
||||||
pub fn serialize_to_xml_model(response: &SerializeResponse, redactions: &RedactionMap) -> String {
|
pub fn serialize_to_xml_model(response: &SerializeResponse, redactions: &RedactionMap) -> String {
|
||||||
let mut dom = rbx_binary::from_reader(response.model_contents.as_slice()).unwrap();
|
let model_content = data_encoding::BASE64
|
||||||
|
.decode(response.model_contents.model().as_bytes())
|
||||||
|
.unwrap();
|
||||||
|
|
||||||
|
let mut dom = rbx_binary::from_reader(model_content.as_slice()).unwrap();
|
||||||
// This makes me realize that maybe we need a `descendants_mut` iter.
|
// This makes me realize that maybe we need a `descendants_mut` iter.
|
||||||
let ref_list: Vec<Ref> = dom.descendants().map(|inst| inst.referent()).collect();
|
let ref_list: Vec<Ref> = dom.descendants().map(|inst| inst.referent()).collect();
|
||||||
for referent in ref_list {
|
for referent in ref_list {
|
||||||
|
|||||||
@@ -41,7 +41,6 @@ gen_build_tests! {
|
|||||||
issue_546,
|
issue_546,
|
||||||
json_as_lua,
|
json_as_lua,
|
||||||
json_model_in_folder,
|
json_model_in_folder,
|
||||||
json_model_legacy_name,
|
|
||||||
module_in_folder,
|
module_in_folder,
|
||||||
module_init,
|
module_init,
|
||||||
nested_runcontext,
|
nested_runcontext,
|
||||||
@@ -55,6 +54,8 @@ gen_build_tests! {
|
|||||||
script_meta_disabled,
|
script_meta_disabled,
|
||||||
server_in_folder,
|
server_in_folder,
|
||||||
server_init,
|
server_init,
|
||||||
|
slugified_name_roundtrip,
|
||||||
|
model_json_name_input,
|
||||||
txt,
|
txt,
|
||||||
txt_in_folder,
|
txt_in_folder,
|
||||||
unresolved_values,
|
unresolved_values,
|
||||||
|
|||||||
@@ -86,4 +86,9 @@ syncback_tests! {
|
|||||||
sync_rules => ["src/module.modulescript", "src/text.text"],
|
sync_rules => ["src/module.modulescript", "src/text.text"],
|
||||||
// Ensures that the `syncUnscriptable` setting works
|
// Ensures that the `syncUnscriptable` setting works
|
||||||
unscriptable_properties => ["default.project.json"],
|
unscriptable_properties => ["default.project.json"],
|
||||||
|
// Ensures that instances with names containing illegal characters get slugified filenames
|
||||||
|
// and preserve their original names in meta.json without forcing directories for leaf scripts
|
||||||
|
slugified_name => ["src/_Script.meta.json", "src/_Script.server.luau", "src/_Folder/init.meta.json"],
|
||||||
|
// Ensures that .model.json files preserve the name property
|
||||||
|
model_json_name => ["src/foo.model.json"],
|
||||||
}
|
}
|
||||||
|
|||||||
Reference in New Issue
Block a user