title: Proxifier
date: Aug 30, 2022
tags: Writeup 10kCTF_RootMe Web MyChallenges



Difficulty: 499 points | 10 solves

Description: How using several URL parser on the same input could be dangerous?

Author: Me :3

Table of content

🕵️ Recon

Like with simple_login, for this challenge we have access to the code. So, the first step is to take a look into it:

const parseUrl = require("parse-url");
const urlParse = require("url-parse");
const fetch = require("node-fetch");
const express = require("express");
const path = require("path");
const fs = require("fs");

const FLAG_PATH = "*censored*";
const app = express();

var getURL = async (url) => {
  // INIT
  var protocol = urlParse(url)["protocol"];
  var host = urlParse(url)["host"];
  var pathname = urlParse(url)["pathname"];

  if (protocol === "https:" && host.endsWith("")) {
    protocol = parseUrl(url)["protocol"];
    host = new URL(pathname, ``)["host"];
    pathname = new URL(pathname, ``)["pathname"];

    // REMOTE
    if (protocol === "https") {
      try {
        var res = await fetch(`${protocol}://${host}${pathname}`);
        res = await res.text();
        return res
      } catch {
        return "Error fetching data.";

      // LOCAL
    } else if (protocol === "file" && host === "") {
      try {
        return fs.readFileSync(pathname, "utf8");
      } catch {
        return "No such file or directory.";

  } else {
    return "Protocol must be https and host end with ''.";

app.get("/proxy", async (req, res) => {
  if (req.query.url !== undefined) {
    res.send(await getURL(req.query.url));
  } else {
    res.send("[ERROR] No URL send!")

app.listen(3000, () => {
  console.log(`Express server listening on port 3000`)

As we can see, the website has an endpoint /proxy which take as an input the url get parameter. Then, our input is sent to the getURL function which is where all the magic occurs. This function will do the following:


🥪 Protocol Confusion

As we can see with the above diagram, if parse-url@7.0.2 return the file:// wrapper we have a chance to get an LFI. But this could be really weird as the first parser must return https which means that 2 parser need to return a different protocol value...


At this point there is two way to find this vulnerability: trying random input into both parser and wait to have the desire output or dive into parse-url@7.0.2 source code and try to find the issue. In this writeup, we will focus on the second one. By going to npm, we could have this github repository: link.

On this, we can get the source code which is really really short. (87 lines)

// Dependencies
import parsePath from "parse-path";
import normalizeUrl from "normalize-url";


const parseUrl = (url, normalize = false) => {


  if (typeof url !== "string" || !url.trim()) {
      throwErr("Invalid url.")

  if (url.length > parseUrl.MAX_INPUT_LENGTH) {
      throwErr("Input exceeds maximum length. If needed, change the value of parseUrl.MAX_INPUT_LENGTH.")

  if (normalize) {
      if (typeof normalize !== "object") {
          normalize = {
              stripHash: false
      url = normalizeUrl(url, normalize)

  const parsed = parsePath(url)


  return parsed;

Inside it, we could see that the most important part of the parsing is done by the function parsePath(url). Looking into the code, we could find that it comes from another library (parse-path) developed by the same person. Thus, let's read his code! (source)


function parsePath(url) {


  try {
      const parsed = new URL(url)
      output.protocols = protocols(parsed)
      output.protocol = output.protocols[0]
      output.port = parsed.port
      output.resource =  parsed.hostname =
      output.user = parsed.username || ""
      output.password = parsed.password || ""
      output.pathname = parsed.pathname
      output.hash = parsed.hash.slice(1) =
      output.href = parsed.href
      output.query = Object.fromEntries(parsed.searchParams)
  } catch (e) {
      // TODO Maybe check if it is a valid local file path
      //      In any case, these will be parsed by higher
      //      level parsers such as parse-url, git-url-parse, git-up
      output.protocols = ["file"]
      output.protocol = output.protocols[0]
      output.port = ""
      output.resource =  ""
      output.user = ""
      output.pathname = ""
      output.hash = "" = ""
      output.href = url
      output.query = {}
      output.parse_failed = true

  return output;

And this is where is starting to be interesting. As we can see, if our input crash during the new URL parsing, it will return the file:// wrapper by default! Let's try this with a simple invalid URL:

const parseUrl = require("parse-url");



  "protocols": [ "file" ],
  "protocol": "file",
  "port": "",
  "resource": "",
  "user": "",
  "password": "",
  "pathname": "",
  "hash": "",
  "search": "",
  "href": "://",
  "query": {}

As we can see, we manage to get the file:// wrapper.


The next step is then to have https:// with url-parse and file:// with parse-url. By looking to some reports on url-parse, we could learn that the library assume that it can return invalid host.

Thus, we know that url-parse allows invalid host and parse-url return file:// wrapper for invalid URL. Combinating both information, we could quickly find a payload that valid all the checks:

PS: This works only before 7.0.2 as there is no check of parse_failed in the main library script.

🏠 Host Confusion

Well, now that we have a bypass for the protocol part, we need to find a way to go around the host check. This one is done into:

host = new URL(pathname, ``)["host"];

Reading the MDN documentation about this API (source), we could learn that depending of the input, the host could change. (example)

The problem here is that there is no example with a path in the first argument which leads to this behavior... So, it is impossible? Obviously not, trying the well-known tricks // and we get exactly what we want! 🎉

new URL("//", ``)


hash: ""
host: ""
hostname: ""
href: ""
origin: ""
password: ""
pathname: "/a"
port: ""
protocol: "https:"
search: ""
URLSearchParams {}
username: ""


Now that we have everything, let's try to get /etc/passwd:


systemd-timesync:x:104:105:systemd Time Synchronization,,,:/run/systemd:/usr/sbin/nologin

Perfect! It works! Now, we need to get the flag and if we remember well the content of the challenge script, it contains this path:

const FLAG_PATH = "*censored*";

So, we need to find a way to leak its name. To do so, we will use the proc linux trick: which returns the executable path of the current proccess (the challenge in the LFI context)


Now we can get the challenge file:

const FLAG_PATH = "a49b4e26e4b6b4638f225fb342a645ce/flag.txt";

🎉 Flag

The last step is to get the flag: :D

Flag: RM{T4k3_C4R3_0f_Y0uR_URL_P4rs3R} 🎉