Exclude/delete webrobots RBI file.

See https://github.com/Homebrew/brew/pull/12135#issuecomment-930114104
This commit is contained in:
Mike McQuaid 2021-09-29 13:42:27 +01:00
parent 0109c699bc
commit 4107380062
No known key found for this signature in database
GPG Key ID: 3338A31AFDB1D829
2 changed files with 1 additions and 170 deletions

View File

@ -50,6 +50,7 @@ module Homebrew
if args.update?
excluded_gems = [
"did_you_mean", # RBI file is already provided by Sorbet
"webrobots", # RBI file is bugged
]
ohai "Updating Tapioca RBI files..."

View File

@ -1,170 +0,0 @@
# DO NOT EDIT MANUALLY
# This is an autogenerated file for types exported from the `webrobots` gem.
# Please instead update this file by running `bin/tapioca gem webrobots`.
# typed: true
class Nokogiri::HTML::Document < ::Nokogiri::XML::Document
def fragment(tags = T.unsafe(nil)); end
def meta_encoding; end
def meta_encoding=(encoding); end
def meta_robots(custom_name = T.unsafe(nil)); end
def nofollow?(custom_name = T.unsafe(nil)); end
def noindex?(custom_name = T.unsafe(nil)); end
def serialize(options = T.unsafe(nil)); end
def title; end
def title=(text); end
def type; end
private
def meta_content_type; end
def parse_meta_robots(custom_name); end
def set_metadata_element(element); end
class << self
def new(*_arg0); end
def parse(string_or_io, url = T.unsafe(nil), encoding = T.unsafe(nil), options = T.unsafe(nil)); end
def read_io(_arg0, _arg1, _arg2, _arg3); end
def read_memory(_arg0, _arg1, _arg2, _arg3); end
end
end
class WebRobots
def initialize(user_agent, options = T.unsafe(nil)); end
def allowed?(url); end
def crawl_delay(url); end
def create_cache; end
def disallowed?(url); end
def error(url); end
def error!(url); end
def flush_cache; end
def option(url, token); end
def options(url); end
def reset(url); end
def sitemaps(url); end
def user_agent; end
private
def crawl_delay_handler(delay, last_checked_at); end
def fetch_robots_txt(site); end
def get_robots_txt(site); end
def http_get(uri); end
def robots_txt_for(url); end
def split_uri(url); end
end
class WebRobots::Error < ::StandardError; end
class WebRobots::ParseError < ::WebRobots::Error
def initialize(message, site); end
def site; end
def to_s; end
end
class WebRobots::RobotsTxt
def initialize(site, records, options = T.unsafe(nil)); end
def allow?(request_uri, user_agent = T.unsafe(nil)); end
def crawl_delay(user_agent = T.unsafe(nil)); end
def error; end
def error!; end
def error=(_arg0); end
def options(user_agent = T.unsafe(nil)); end
def site; end
def sitemaps; end
def timestamp; end
private
def find_record(user_agent = T.unsafe(nil)); end
def target(user_agent = T.unsafe(nil)); end
class << self
def unfetchable(site, reason, target = T.unsafe(nil)); end
end
end
class WebRobots::RobotsTxt::AccessControlLine < ::WebRobots::RobotsTxt::Line
def compile; end
def match?(request_uri); end
end
class WebRobots::RobotsTxt::AgentLine < ::WebRobots::RobotsTxt::Line
def compile; end
def pattern; end
end
class WebRobots::RobotsTxt::AllowLine < ::WebRobots::RobotsTxt::AccessControlLine
def allow?; end
end
class WebRobots::RobotsTxt::CrawlDelayLine < ::WebRobots::RobotsTxt::Line
def compile; end
def delay; end
end
WebRobots::RobotsTxt::DISALLOW_ALL = T.let(T.unsafe(nil), String)
class WebRobots::RobotsTxt::DisallowLine < ::WebRobots::RobotsTxt::AccessControlLine
def allow?; end
end
class WebRobots::RobotsTxt::ExtentionLine < ::WebRobots::RobotsTxt::Line; end
class WebRobots::RobotsTxt::Line
def initialize(token, value); end
def compile; end
def token; end
def value; end
end
class WebRobots::RobotsTxt::Parser < ::Racc::Parser
def initialize(target, crawl_delay_handler = T.unsafe(nil)); end
def _reduce_1(val, _values, result); end
def _reduce_17(val, _values, result); end
def _reduce_18(val, _values, result); end
def _reduce_19(val, _values, result); end
def _reduce_2(val, _values, result); end
def _reduce_20(val, _values, result); end
def _reduce_21(val, _values, result); end
def _reduce_24(val, _values, result); end
def _reduce_25(val, _values, result); end
def _reduce_26(val, _values, result); end
def _reduce_28(val, _values, result); end
def _reduce_31(val, _values, result); end
def _reduce_32(val, _values, result); end
def _reduce_38(val, _values, result); end
def _reduce_39(val, _values, result); end
def _reduce_40(val, _values, result); end
def _reduce_41(val, _values, result); end
def _reduce_none(val, _values, result); end
def next_token; end
def on_error(token_id, value, stack); end
def parse(input, site); end
def parse!(input, site); end
def parse_error(message); end
end
WebRobots::RobotsTxt::Parser::KNOWN_TOKENS = T.let(T.unsafe(nil), Array)
WebRobots::RobotsTxt::Parser::RE_KNOWN_TOKENS = T.let(T.unsafe(nil), Regexp)
WebRobots::RobotsTxt::Parser::Racc_arg = T.let(T.unsafe(nil), Array)
WebRobots::RobotsTxt::Parser::Racc_token_to_s_table = T.let(T.unsafe(nil), Array)
class WebRobots::RobotsTxt::Record
def initialize(agentlines, rulelines); end
def allow?(request_uri); end
def default?; end
def delay; end
def match?(user_agent); end
def options; end
end
module Webrobots; end
Webrobots::VERSION = T.let(T.unsafe(nil), String)