Exclude/delete webrobots RBI file.
See https://github.com/Homebrew/brew/pull/12135#issuecomment-930114104
This commit is contained in:
parent
0109c699bc
commit
4107380062
@ -50,6 +50,7 @@ module Homebrew
|
|||||||
if args.update?
|
if args.update?
|
||||||
excluded_gems = [
|
excluded_gems = [
|
||||||
"did_you_mean", # RBI file is already provided by Sorbet
|
"did_you_mean", # RBI file is already provided by Sorbet
|
||||||
|
"webrobots", # RBI file is bugged
|
||||||
]
|
]
|
||||||
|
|
||||||
ohai "Updating Tapioca RBI files..."
|
ohai "Updating Tapioca RBI files..."
|
||||||
|
@ -1,170 +0,0 @@
|
|||||||
# DO NOT EDIT MANUALLY
|
|
||||||
# This is an autogenerated file for types exported from the `webrobots` gem.
|
|
||||||
# Please instead update this file by running `bin/tapioca gem webrobots`.
|
|
||||||
|
|
||||||
# typed: true
|
|
||||||
|
|
||||||
class Nokogiri::HTML::Document < ::Nokogiri::XML::Document
|
|
||||||
def fragment(tags = T.unsafe(nil)); end
|
|
||||||
def meta_encoding; end
|
|
||||||
def meta_encoding=(encoding); end
|
|
||||||
def meta_robots(custom_name = T.unsafe(nil)); end
|
|
||||||
def nofollow?(custom_name = T.unsafe(nil)); end
|
|
||||||
def noindex?(custom_name = T.unsafe(nil)); end
|
|
||||||
def serialize(options = T.unsafe(nil)); end
|
|
||||||
def title; end
|
|
||||||
def title=(text); end
|
|
||||||
def type; end
|
|
||||||
|
|
||||||
private
|
|
||||||
|
|
||||||
def meta_content_type; end
|
|
||||||
def parse_meta_robots(custom_name); end
|
|
||||||
def set_metadata_element(element); end
|
|
||||||
|
|
||||||
class << self
|
|
||||||
def new(*_arg0); end
|
|
||||||
def parse(string_or_io, url = T.unsafe(nil), encoding = T.unsafe(nil), options = T.unsafe(nil)); end
|
|
||||||
def read_io(_arg0, _arg1, _arg2, _arg3); end
|
|
||||||
def read_memory(_arg0, _arg1, _arg2, _arg3); end
|
|
||||||
end
|
|
||||||
end
|
|
||||||
|
|
||||||
class WebRobots
|
|
||||||
def initialize(user_agent, options = T.unsafe(nil)); end
|
|
||||||
|
|
||||||
def allowed?(url); end
|
|
||||||
def crawl_delay(url); end
|
|
||||||
def create_cache; end
|
|
||||||
def disallowed?(url); end
|
|
||||||
def error(url); end
|
|
||||||
def error!(url); end
|
|
||||||
def flush_cache; end
|
|
||||||
def option(url, token); end
|
|
||||||
def options(url); end
|
|
||||||
def reset(url); end
|
|
||||||
def sitemaps(url); end
|
|
||||||
def user_agent; end
|
|
||||||
|
|
||||||
private
|
|
||||||
|
|
||||||
def crawl_delay_handler(delay, last_checked_at); end
|
|
||||||
def fetch_robots_txt(site); end
|
|
||||||
def get_robots_txt(site); end
|
|
||||||
def http_get(uri); end
|
|
||||||
def robots_txt_for(url); end
|
|
||||||
def split_uri(url); end
|
|
||||||
end
|
|
||||||
|
|
||||||
class WebRobots::Error < ::StandardError; end
|
|
||||||
|
|
||||||
class WebRobots::ParseError < ::WebRobots::Error
|
|
||||||
def initialize(message, site); end
|
|
||||||
|
|
||||||
def site; end
|
|
||||||
def to_s; end
|
|
||||||
end
|
|
||||||
|
|
||||||
class WebRobots::RobotsTxt
|
|
||||||
def initialize(site, records, options = T.unsafe(nil)); end
|
|
||||||
|
|
||||||
def allow?(request_uri, user_agent = T.unsafe(nil)); end
|
|
||||||
def crawl_delay(user_agent = T.unsafe(nil)); end
|
|
||||||
def error; end
|
|
||||||
def error!; end
|
|
||||||
def error=(_arg0); end
|
|
||||||
def options(user_agent = T.unsafe(nil)); end
|
|
||||||
def site; end
|
|
||||||
def sitemaps; end
|
|
||||||
def timestamp; end
|
|
||||||
|
|
||||||
private
|
|
||||||
|
|
||||||
def find_record(user_agent = T.unsafe(nil)); end
|
|
||||||
def target(user_agent = T.unsafe(nil)); end
|
|
||||||
|
|
||||||
class << self
|
|
||||||
def unfetchable(site, reason, target = T.unsafe(nil)); end
|
|
||||||
end
|
|
||||||
end
|
|
||||||
|
|
||||||
class WebRobots::RobotsTxt::AccessControlLine < ::WebRobots::RobotsTxt::Line
|
|
||||||
def compile; end
|
|
||||||
def match?(request_uri); end
|
|
||||||
end
|
|
||||||
|
|
||||||
class WebRobots::RobotsTxt::AgentLine < ::WebRobots::RobotsTxt::Line
|
|
||||||
def compile; end
|
|
||||||
def pattern; end
|
|
||||||
end
|
|
||||||
|
|
||||||
class WebRobots::RobotsTxt::AllowLine < ::WebRobots::RobotsTxt::AccessControlLine
|
|
||||||
def allow?; end
|
|
||||||
end
|
|
||||||
|
|
||||||
class WebRobots::RobotsTxt::CrawlDelayLine < ::WebRobots::RobotsTxt::Line
|
|
||||||
def compile; end
|
|
||||||
def delay; end
|
|
||||||
end
|
|
||||||
|
|
||||||
WebRobots::RobotsTxt::DISALLOW_ALL = T.let(T.unsafe(nil), String)
|
|
||||||
|
|
||||||
class WebRobots::RobotsTxt::DisallowLine < ::WebRobots::RobotsTxt::AccessControlLine
|
|
||||||
def allow?; end
|
|
||||||
end
|
|
||||||
|
|
||||||
class WebRobots::RobotsTxt::ExtentionLine < ::WebRobots::RobotsTxt::Line; end
|
|
||||||
|
|
||||||
class WebRobots::RobotsTxt::Line
|
|
||||||
def initialize(token, value); end
|
|
||||||
|
|
||||||
def compile; end
|
|
||||||
def token; end
|
|
||||||
def value; end
|
|
||||||
end
|
|
||||||
|
|
||||||
class WebRobots::RobotsTxt::Parser < ::Racc::Parser
|
|
||||||
def initialize(target, crawl_delay_handler = T.unsafe(nil)); end
|
|
||||||
|
|
||||||
def _reduce_1(val, _values, result); end
|
|
||||||
def _reduce_17(val, _values, result); end
|
|
||||||
def _reduce_18(val, _values, result); end
|
|
||||||
def _reduce_19(val, _values, result); end
|
|
||||||
def _reduce_2(val, _values, result); end
|
|
||||||
def _reduce_20(val, _values, result); end
|
|
||||||
def _reduce_21(val, _values, result); end
|
|
||||||
def _reduce_24(val, _values, result); end
|
|
||||||
def _reduce_25(val, _values, result); end
|
|
||||||
def _reduce_26(val, _values, result); end
|
|
||||||
def _reduce_28(val, _values, result); end
|
|
||||||
def _reduce_31(val, _values, result); end
|
|
||||||
def _reduce_32(val, _values, result); end
|
|
||||||
def _reduce_38(val, _values, result); end
|
|
||||||
def _reduce_39(val, _values, result); end
|
|
||||||
def _reduce_40(val, _values, result); end
|
|
||||||
def _reduce_41(val, _values, result); end
|
|
||||||
def _reduce_none(val, _values, result); end
|
|
||||||
def next_token; end
|
|
||||||
def on_error(token_id, value, stack); end
|
|
||||||
def parse(input, site); end
|
|
||||||
def parse!(input, site); end
|
|
||||||
def parse_error(message); end
|
|
||||||
end
|
|
||||||
|
|
||||||
WebRobots::RobotsTxt::Parser::KNOWN_TOKENS = T.let(T.unsafe(nil), Array)
|
|
||||||
WebRobots::RobotsTxt::Parser::RE_KNOWN_TOKENS = T.let(T.unsafe(nil), Regexp)
|
|
||||||
WebRobots::RobotsTxt::Parser::Racc_arg = T.let(T.unsafe(nil), Array)
|
|
||||||
WebRobots::RobotsTxt::Parser::Racc_token_to_s_table = T.let(T.unsafe(nil), Array)
|
|
||||||
|
|
||||||
class WebRobots::RobotsTxt::Record
|
|
||||||
def initialize(agentlines, rulelines); end
|
|
||||||
|
|
||||||
def allow?(request_uri); end
|
|
||||||
def default?; end
|
|
||||||
def delay; end
|
|
||||||
def match?(user_agent); end
|
|
||||||
def options; end
|
|
||||||
end
|
|
||||||
|
|
||||||
module Webrobots; end
|
|
||||||
Webrobots::VERSION = T.let(T.unsafe(nil), String)
|
|
Loading…
x
Reference in New Issue
Block a user