From 410738006266824b59ec844b67b140fb2cccfcfc Mon Sep 17 00:00:00 2001 From: Mike McQuaid Date: Wed, 29 Sep 2021 13:42:27 +0100 Subject: [PATCH] Exclude/delete webrobots RBI file. See https://github.com/Homebrew/brew/pull/12135#issuecomment-930114104 --- Library/Homebrew/dev-cmd/typecheck.rb | 1 + .../sorbet/rbi/gems/webrobots@0.1.2.rbi | 170 ------------------ 2 files changed, 1 insertion(+), 170 deletions(-) delete mode 100644 Library/Homebrew/sorbet/rbi/gems/webrobots@0.1.2.rbi diff --git a/Library/Homebrew/dev-cmd/typecheck.rb b/Library/Homebrew/dev-cmd/typecheck.rb index a178281649..7f1c41e541 100644 --- a/Library/Homebrew/dev-cmd/typecheck.rb +++ b/Library/Homebrew/dev-cmd/typecheck.rb @@ -50,6 +50,7 @@ module Homebrew if args.update? excluded_gems = [ "did_you_mean", # RBI file is already provided by Sorbet + "webrobots", # RBI file is bugged ] ohai "Updating Tapioca RBI files..." diff --git a/Library/Homebrew/sorbet/rbi/gems/webrobots@0.1.2.rbi b/Library/Homebrew/sorbet/rbi/gems/webrobots@0.1.2.rbi deleted file mode 100644 index 5ea05a93c2..0000000000 --- a/Library/Homebrew/sorbet/rbi/gems/webrobots@0.1.2.rbi +++ /dev/null @@ -1,170 +0,0 @@ -# DO NOT EDIT MANUALLY -# This is an autogenerated file for types exported from the `webrobots` gem. -# Please instead update this file by running `bin/tapioca gem webrobots`. - -# typed: true - -class Nokogiri::HTML::Document < ::Nokogiri::XML::Document - def fragment(tags = T.unsafe(nil)); end - def meta_encoding; end - def meta_encoding=(encoding); end - def meta_robots(custom_name = T.unsafe(nil)); end - def nofollow?(custom_name = T.unsafe(nil)); end - def noindex?(custom_name = T.unsafe(nil)); end - def serialize(options = T.unsafe(nil)); end - def title; end - def title=(text); end - def type; end - - private - - def meta_content_type; end - def parse_meta_robots(custom_name); end - def set_metadata_element(element); end - - class << self - def new(*_arg0); end - def parse(string_or_io, url = T.unsafe(nil), encoding = T.unsafe(nil), options = T.unsafe(nil)); end - def read_io(_arg0, _arg1, _arg2, _arg3); end - def read_memory(_arg0, _arg1, _arg2, _arg3); end - end -end - -class WebRobots - def initialize(user_agent, options = T.unsafe(nil)); end - - def allowed?(url); end - def crawl_delay(url); end - def create_cache; end - def disallowed?(url); end - def error(url); end - def error!(url); end - def flush_cache; end - def option(url, token); end - def options(url); end - def reset(url); end - def sitemaps(url); end - def user_agent; end - - private - - def crawl_delay_handler(delay, last_checked_at); end - def fetch_robots_txt(site); end - def get_robots_txt(site); end - def http_get(uri); end - def robots_txt_for(url); end - def split_uri(url); end -end - -class WebRobots::Error < ::StandardError; end - -class WebRobots::ParseError < ::WebRobots::Error - def initialize(message, site); end - - def site; end - def to_s; end -end - -class WebRobots::RobotsTxt - def initialize(site, records, options = T.unsafe(nil)); end - - def allow?(request_uri, user_agent = T.unsafe(nil)); end - def crawl_delay(user_agent = T.unsafe(nil)); end - def error; end - def error!; end - def error=(_arg0); end - def options(user_agent = T.unsafe(nil)); end - def site; end - def sitemaps; end - def timestamp; end - - private - - def find_record(user_agent = T.unsafe(nil)); end - def target(user_agent = T.unsafe(nil)); end - - class << self - def unfetchable(site, reason, target = T.unsafe(nil)); end - end -end - -class WebRobots::RobotsTxt::AccessControlLine < ::WebRobots::RobotsTxt::Line - def compile; end - def match?(request_uri); end -end - -class WebRobots::RobotsTxt::AgentLine < ::WebRobots::RobotsTxt::Line - def compile; end - def pattern; end -end - -class WebRobots::RobotsTxt::AllowLine < ::WebRobots::RobotsTxt::AccessControlLine - def allow?; end -end - -class WebRobots::RobotsTxt::CrawlDelayLine < ::WebRobots::RobotsTxt::Line - def compile; end - def delay; end -end - -WebRobots::RobotsTxt::DISALLOW_ALL = T.let(T.unsafe(nil), String) - -class WebRobots::RobotsTxt::DisallowLine < ::WebRobots::RobotsTxt::AccessControlLine - def allow?; end -end - -class WebRobots::RobotsTxt::ExtentionLine < ::WebRobots::RobotsTxt::Line; end - -class WebRobots::RobotsTxt::Line - def initialize(token, value); end - - def compile; end - def token; end - def value; end -end - -class WebRobots::RobotsTxt::Parser < ::Racc::Parser - def initialize(target, crawl_delay_handler = T.unsafe(nil)); end - - def _reduce_1(val, _values, result); end - def _reduce_17(val, _values, result); end - def _reduce_18(val, _values, result); end - def _reduce_19(val, _values, result); end - def _reduce_2(val, _values, result); end - def _reduce_20(val, _values, result); end - def _reduce_21(val, _values, result); end - def _reduce_24(val, _values, result); end - def _reduce_25(val, _values, result); end - def _reduce_26(val, _values, result); end - def _reduce_28(val, _values, result); end - def _reduce_31(val, _values, result); end - def _reduce_32(val, _values, result); end - def _reduce_38(val, _values, result); end - def _reduce_39(val, _values, result); end - def _reduce_40(val, _values, result); end - def _reduce_41(val, _values, result); end - def _reduce_none(val, _values, result); end - def next_token; end - def on_error(token_id, value, stack); end - def parse(input, site); end - def parse!(input, site); end - def parse_error(message); end -end - -WebRobots::RobotsTxt::Parser::KNOWN_TOKENS = T.let(T.unsafe(nil), Array) -WebRobots::RobotsTxt::Parser::RE_KNOWN_TOKENS = T.let(T.unsafe(nil), Regexp) -WebRobots::RobotsTxt::Parser::Racc_arg = T.let(T.unsafe(nil), Array) -WebRobots::RobotsTxt::Parser::Racc_token_to_s_table = T.let(T.unsafe(nil), Array) - -class WebRobots::RobotsTxt::Record - def initialize(agentlines, rulelines); end - - def allow?(request_uri); end - def default?; end - def delay; end - def match?(user_agent); end - def options; end -end - -module Webrobots; end -Webrobots::VERSION = T.let(T.unsafe(nil), String)