mirror of
https://github.com/open-goal/jak-project.git
synced 2024-10-20 11:26:18 -04:00
c162c66118
This PR does two main things: 1. Work through the main low-hanging fruit issues in the formatter keeping it from feeling mature and usable 2. Iterate and prove that point by formatting all of the Jak 1 code base. **This has removed around 100K lines in total.** - The decompiler will now format it's results for jak 1 to keep things from drifting back to where they were. This is controlled by a new config flag `format_code`. How am I confident this hasn't broken anything?: - I compiled the entire project and stored it's `out/jak1/obj` files separately - I then recompiled the project after formatting and wrote a script that md5's each file and compares it (`compare-compilation-outputs.py` - The results (eventually) were the same: ![Screenshot 2024-05-25 132900](https://github.com/open-goal/jak-project/assets/13153231/015e6f20-8d19-49b7-9951-97fa88ddc6c2) > This proves that the only difference before and after is non-critical whitespace for all code/macros that is actually in use. I'm still aware of improvements that could be made to the formatter, as well as general optimization of it's performance. But in general these are for rare or non-critical situations in my opinion and I'll work through them before doing Jak 2. The vast majority looks great and is working properly at this point. Those known issues are the following if you are curious: ![image](https://github.com/open-goal/jak-project/assets/13153231/0edfaba1-6d36-40f5-ab23-0642209867c4)
166 lines
4.7 KiB
C++
166 lines
4.7 KiB
C++
// TODO - eventually replace our `goalc` tests with this setup
|
|
// A simple test runner framework for debugging / iterating on the formatter
|
|
// Tests are defined in files as such:
|
|
|
|
/*
|
|
===
|
|
TEST NAME
|
|
===
|
|
|
|
INPUT
|
|
|
|
---
|
|
|
|
EXPECTED OUTPUT
|
|
|
|
*/
|
|
|
|
// Test files can contain multiple tests, upon running we will recurse a directory
|
|
// looking for any `.test` files and run them through the framework
|
|
//
|
|
// Any differences will be diff'd and displayed
|
|
|
|
#include "common/formatter/formatter.h"
|
|
#include "common/util/FileUtil.h"
|
|
#include "common/util/string_util.h"
|
|
|
|
#include "gtest/gtest.h"
|
|
|
|
#include "fmt/color.h"
|
|
#include "fmt/core.h"
|
|
|
|
struct TestDefinition {
|
|
std::string name;
|
|
std::string input;
|
|
std::string output;
|
|
};
|
|
|
|
std::vector<TestDefinition> get_test_definitions(const fs::path& file_path) {
|
|
std::vector<TestDefinition> tests;
|
|
// Read in the file, and run the test
|
|
const auto contents = str_util::split(file_util::read_text_file(file_path));
|
|
if (contents.empty() || (contents.size() == 1 && contents.at(0).empty())) {
|
|
return tests;
|
|
}
|
|
TestDefinition curr_test;
|
|
size_t i = 0;
|
|
while (i < contents.size()) {
|
|
const auto& line = contents.at(i);
|
|
if (line == "===") {
|
|
curr_test = TestDefinition();
|
|
curr_test.name = contents.at(i + 1);
|
|
i += 3;
|
|
continue;
|
|
}
|
|
// Parse the input and output
|
|
if (!curr_test.name.empty() && line.empty()) {
|
|
i++;
|
|
while (true) {
|
|
if (str_util::trim(contents.at(i)) == "---") {
|
|
i++;
|
|
curr_test.input = str_util::trim(curr_test.input);
|
|
break;
|
|
}
|
|
curr_test.input += contents.at(i) + "\n";
|
|
i++;
|
|
}
|
|
i++;
|
|
while (true) {
|
|
if (i == contents.size() || contents.at(i) == "===") {
|
|
curr_test.output = str_util::trim(curr_test.output);
|
|
tests.push_back(curr_test);
|
|
break;
|
|
}
|
|
curr_test.output += contents.at(i) + "\n";
|
|
i++;
|
|
}
|
|
continue;
|
|
}
|
|
}
|
|
return tests;
|
|
}
|
|
|
|
bool has_important_tests(const fs::path& file_path) {
|
|
const auto& tests = get_test_definitions(file_path);
|
|
for (const auto& test : tests) {
|
|
if (str_util::starts_with(test.name, "!")) {
|
|
return true;
|
|
}
|
|
}
|
|
return false;
|
|
}
|
|
|
|
// TODO - consider adding a test that auto-formats all of goal_src (there should be no errors)
|
|
|
|
bool run_tests(const fs::path& file_path, const bool only_important_tests) {
|
|
const auto& tests = get_test_definitions(file_path);
|
|
// Run the tests, report successes and failures
|
|
bool test_failed = false;
|
|
fmt::print("{}:\n", fmt::styled(file_util::base_name(file_path.string()),
|
|
fmt::emphasis::bold | fg(fmt::color::cyan)));
|
|
for (const auto& test : tests) {
|
|
if (only_important_tests && !str_util::starts_with(test.name, "!")) {
|
|
continue;
|
|
}
|
|
if (str_util::contains(test.name, "TODO")) {
|
|
// ignore the output
|
|
fmt::print(" ⚠️ - {}\n", test.name);
|
|
continue;
|
|
}
|
|
const auto formatted_result = formatter::format_code(test.input);
|
|
if (formatted_result && str_util::starts_with(test.name, "!?")) {
|
|
fmt::print("FORMATTED RESULT:\n\n{}\n\n", formatted_result.value());
|
|
}
|
|
if (!formatted_result) {
|
|
// Unable to parse, was that expected?
|
|
if (test.output == "__THROWS__") {
|
|
fmt::print(" ✅ - {}\n", test.name);
|
|
} else {
|
|
fmt::print(" ❌ - {}\n", test.name);
|
|
fmt::print("Unable to Format\n");
|
|
test_failed = true;
|
|
}
|
|
} else if (formatted_result != test.output) {
|
|
fmt::print(" ❌ - {}\n", test.name);
|
|
fmt::print("{}\n", str_util::diff(test.output, formatted_result.value()));
|
|
test_failed = true;
|
|
} else {
|
|
fmt::print(" ✅ - {}\n", test.name);
|
|
}
|
|
}
|
|
return test_failed;
|
|
}
|
|
|
|
bool find_and_run_tests() {
|
|
// TODO - fails when it finds no tests
|
|
try {
|
|
// Enumerate test files
|
|
const auto test_files = file_util::find_files_recursively(
|
|
file_util::get_file_path({"test/common/formatter/corpus"}), std::regex("^.*\\.test.gc$"));
|
|
bool failed = false;
|
|
// First do a pass to see if any tests are meant to be prioritized for debugging
|
|
bool only_important_tests = false;
|
|
for (const auto& file : test_files) {
|
|
only_important_tests = has_important_tests(file);
|
|
if (only_important_tests) {
|
|
break;
|
|
}
|
|
}
|
|
for (const auto& file : test_files) {
|
|
// don't fail fast, but any failure means we return false
|
|
if (failed) {
|
|
run_tests(file, only_important_tests);
|
|
} else {
|
|
failed = run_tests(file, only_important_tests);
|
|
}
|
|
}
|
|
return !failed;
|
|
} catch (std::exception& e) {
|
|
return false;
|
|
}
|
|
}
|
|
|
|
TEST(Formatter, FormatterTests) {
|
|
EXPECT_TRUE(find_and_run_tests());
|
|
}
|