librustdoc: remove unnecessary `as_slice()` calls

This commit is contained in:
Jorge Aparicio 2014-11-27 14:19:27 -05:00
parent 8bb5ef9df5
commit e6bd217ce8
7 changed files with 39 additions and 39 deletions

View File

@ -341,10 +341,10 @@ fn build_impl(cx: &DocContext, tcx: &ty::ctxt,
fn is_doc_hidden(a: &clean::Attribute) -> bool {
match *a {
clean::List(ref name, ref inner) if name.as_slice() == "doc" => {
clean::List(ref name, ref inner) if *name == "doc" => {
inner.iter().any(|a| {
match *a {
clean::Word(ref s) => s.as_slice() == "hidden",
clean::Word(ref s) => *s == "hidden",
_ => false,
}
})

View File

@ -256,7 +256,7 @@ impl Item {
pub fn doc_list<'a>(&'a self) -> Option<&'a [Attribute]> {
for attr in self.attrs.iter() {
match *attr {
List(ref x, ref list) if "doc" == x.as_slice() => {
List(ref x, ref list) if "doc" == *x => {
return Some(list.as_slice());
}
_ => {}
@ -270,7 +270,7 @@ impl Item {
pub fn doc_value<'a>(&'a self) -> Option<&'a str> {
for attr in self.attrs.iter() {
match *attr {
NameValue(ref x, ref v) if "doc" == x.as_slice() => {
NameValue(ref x, ref v) if "doc" == *x => {
return Some(v.as_slice());
}
_ => {}
@ -284,7 +284,7 @@ impl Item {
Some(ref l) => {
for innerattr in l.iter() {
match *innerattr {
Word(ref s) if "hidden" == s.as_slice() => {
Word(ref s) if "hidden" == *s => {
return true
}
_ => (),
@ -1217,13 +1217,13 @@ impl PrimitiveType {
fn find(attrs: &[Attribute]) -> Option<PrimitiveType> {
for attr in attrs.iter() {
let list = match *attr {
List(ref k, ref l) if k.as_slice() == "doc" => l,
List(ref k, ref l) if *k == "doc" => l,
_ => continue,
};
for sub_attr in list.iter() {
let value = match *sub_attr {
NameValue(ref k, ref v)
if k.as_slice() == "primitive" => v.as_slice(),
if *k == "primitive" => v.as_slice(),
_ => continue,
};
match PrimitiveType::from_str(value) {

View File

@ -251,8 +251,8 @@ fn path(w: &mut fmt::Formatter, path: &clean::Path, print_all: bool,
Some(root) => {
let mut root = String::from_str(root.as_slice());
for seg in path.segments[..amt].iter() {
if "super" == seg.name.as_slice() ||
"self" == seg.name.as_slice() {
if "super" == seg.name ||
"self" == seg.name {
try!(write!(w, "{}::", seg.name));
} else {
root.push_str(seg.name.as_slice());
@ -337,7 +337,7 @@ fn primitive_link(f: &mut fmt::Formatter,
Some(root) => {
try!(write!(f, "<a href='{}{}/primitive.{}.html'>",
root,
path.ref0().as_slice().head().unwrap(),
path.ref0().head().unwrap(),
prim.to_url_str()));
needs_termination = true;
}

View File

@ -231,7 +231,7 @@ pub fn render(w: &mut fmt::Formatter, s: &str, print_toc: bool) -> fmt::Result {
};
// Transform the contents of the header into a hyphenated string
let id = s.as_slice().words().map(|s| s.to_ascii_lower())
let id = s.words().map(|s| s.to_ascii_lower())
.collect::<Vec<String>>().connect("-");
// This is a terrible hack working around how hoedown gives us rendered
@ -393,7 +393,7 @@ impl LangString {
let mut seen_other_tags = false;
let mut data = LangString::all_false();
let mut tokens = string.as_slice().split(|c: char|
let mut tokens = string.split(|c: char|
!(c == '_' || c == '-' || c.is_alphanumeric())
);

View File

@ -277,15 +277,15 @@ pub fn run(mut krate: clean::Crate,
for attr in attrs.iter() {
match *attr {
clean::NameValue(ref x, ref s)
if "html_favicon_url" == x.as_slice() => {
if "html_favicon_url" == *x => {
cx.layout.favicon = s.to_string();
}
clean::NameValue(ref x, ref s)
if "html_logo_url" == x.as_slice() => {
if "html_logo_url" == *x => {
cx.layout.logo = s.to_string();
}
clean::NameValue(ref x, ref s)
if "html_playground_url" == x.as_slice() => {
if "html_playground_url" == *x => {
cx.layout.playground_url = s.to_string();
markdown::PLAYGROUND_KRATE.with(|slot| {
if slot.borrow().is_none() {
@ -295,7 +295,7 @@ pub fn run(mut krate: clean::Crate,
});
}
clean::Word(ref x)
if "html_no_source" == x.as_slice() => {
if "html_no_source" == *x => {
cx.include_sources = false;
}
_ => {}
@ -434,7 +434,7 @@ fn build_index(krate: &clean::Crate, cache: &mut Cache) -> io::IoResult<String>
for (i, item) in cache.search_index.iter().enumerate() {
// Omit the path if it is same to that of the prior item.
let path;
if lastpath.as_slice() == item.path.as_slice() {
if lastpath == item.path {
path = "";
} else {
lastpath = item.path.to_string();
@ -513,10 +513,10 @@ fn write_shared(cx: &Context,
if path.exists() {
for line in BufferedReader::new(File::open(path)).lines() {
let line = try!(line);
if !line.as_slice().starts_with(key) {
if !line.starts_with(key) {
continue
}
if line.as_slice().starts_with(
if line.starts_with(
format!("{}['{}']", key, krate).as_slice()) {
continue
}
@ -670,12 +670,12 @@ fn extern_location(e: &clean::ExternalCrate, dst: &Path) -> ExternalLocation {
// external crate
for attr in e.attrs.iter() {
match *attr {
clean::List(ref x, ref list) if "doc" == x.as_slice() => {
clean::List(ref x, ref list) if "doc" == *x => {
for attr in list.iter() {
match *attr {
clean::NameValue(ref x, ref s)
if "html_root_url" == x.as_slice() => {
if s.as_slice().ends_with("/") {
if "html_root_url" == *x => {
if s.ends_with("/") {
return Remote(s.to_string());
}
return Remote(format!("{}/", s));
@ -964,7 +964,7 @@ impl DocFolder for Cache {
let dox = match attrs.into_iter().find(|a| {
match *a {
clean::NameValue(ref x, _)
if "doc" == x.as_slice() => {
if "doc" == *x => {
true
}
_ => false

View File

@ -299,7 +299,7 @@ fn acquire_input(input: &str,
fn parse_externs(matches: &getopts::Matches) -> Result<core::Externs, String> {
let mut externs = HashMap::new();
for arg in matches.opt_strs("extern").iter() {
let mut parts = arg.as_slice().splitn(1, '=');
let mut parts = arg.splitn(1, '=');
let name = match parts.next() {
Some(s) => s,
None => {
@ -363,18 +363,18 @@ fn rust_input(cratefile: &str, externs: core::Externs, matches: &getopts::Matche
for inner in nested.iter() {
match *inner {
clean::Word(ref x)
if "no_default_passes" == x.as_slice() => {
if "no_default_passes" == *x => {
default_passes = false;
}
clean::NameValue(ref x, ref value)
if "passes" == x.as_slice() => {
for pass in value.as_slice().words() {
if "passes" == *x => {
for pass in value.words() {
passes.push(pass.to_string());
}
}
clean::NameValue(ref x, ref value)
if "plugins" == x.as_slice() => {
for p in value.as_slice().words() {
if "plugins" == *x => {
for p in value.words() {
plugins.push(p.to_string());
}
}
@ -397,7 +397,7 @@ fn rust_input(cratefile: &str, externs: core::Externs, matches: &getopts::Matche
for pass in passes.iter() {
let plugin = match PASSES.iter()
.position(|&(p, _, _)| {
p == pass.as_slice()
p == *pass
}) {
Some(i) => PASSES[i].val1(),
None => {
@ -434,7 +434,7 @@ fn json_input(input: &str) -> Result<Output, String> {
// Make sure the schema is what we expect
match obj.remove(&"schema".to_string()) {
Some(Json::String(version)) => {
if version.as_slice() != SCHEMA_VERSION {
if version != SCHEMA_VERSION {
return Err(format!(
"sorry, but I only understand version {}",
SCHEMA_VERSION))

View File

@ -258,7 +258,7 @@ pub fn unindent_comments(krate: clean::Crate) -> plugins::PluginResult {
for attr in i.attrs.iter() {
match attr {
&clean::NameValue(ref x, ref s)
if "doc" == x.as_slice() => {
if "doc" == *x => {
avec.push(clean::NameValue("doc".to_string(),
unindent(s.as_slice())))
}
@ -283,7 +283,7 @@ pub fn collapse_docs(krate: clean::Crate) -> plugins::PluginResult {
for attr in i.attrs.iter() {
match *attr {
clean::NameValue(ref x, ref s)
if "doc" == x.as_slice() => {
if "doc" == *x => {
docstr.push_str(s.as_slice());
docstr.push('\n');
},
@ -291,7 +291,7 @@ pub fn collapse_docs(krate: clean::Crate) -> plugins::PluginResult {
}
}
let mut a: Vec<clean::Attribute> = i.attrs.iter().filter(|&a| match a {
&clean::NameValue(ref x, _) if "doc" == x.as_slice() => false,
&clean::NameValue(ref x, _) if "doc" == *x => false,
_ => true
}).map(|x| x.clone()).collect();
if docstr.len() > 0 {
@ -374,14 +374,14 @@ mod unindent_tests {
fn should_unindent() {
let s = " line1\n line2".to_string();
let r = unindent(s.as_slice());
assert_eq!(r.as_slice(), "line1\nline2");
assert_eq!(r, "line1\nline2");
}
#[test]
fn should_unindent_multiple_paragraphs() {
let s = " line1\n\n line2".to_string();
let r = unindent(s.as_slice());
assert_eq!(r.as_slice(), "line1\n\nline2");
assert_eq!(r, "line1\n\nline2");
}
#[test]
@ -390,7 +390,7 @@ mod unindent_tests {
// base indentation and should be preserved
let s = " line1\n\n line2".to_string();
let r = unindent(s.as_slice());
assert_eq!(r.as_slice(), "line1\n\n line2");
assert_eq!(r, "line1\n\n line2");
}
#[test]
@ -402,13 +402,13 @@ mod unindent_tests {
// and continue here"]
let s = "line1\n line2".to_string();
let r = unindent(s.as_slice());
assert_eq!(r.as_slice(), "line1\nline2");
assert_eq!(r, "line1\nline2");
}
#[test]
fn should_not_ignore_first_line_indent_in_a_single_line_para() {
let s = "line1\n\n line2".to_string();
let r = unindent(s.as_slice());
assert_eq!(r.as_slice(), "line1\n\n line2");
assert_eq!(r, "line1\n\n line2");
}
}