Working with Text Data

Series is equipped with a set of string processing methods that make it easy to operate on each element of the array. Perhaps most importantly, these methods exclude missing/NA values automatically. These are accessed via the Series’s str attribute and generally have names matching the equivalent (scalar) built-in string methods:

In [1]: s = Series(['A', 'B', 'C', 'Aaba', 'Baca', np.nan, 'CABA', 'dog', 'cat'])

In [2]: s.str.lower()
0       a
1       b
2       c
3    aaba
4    baca
5     NaN
6    caba
7     dog
8     cat
dtype: object

In [3]: s.str.upper()
0       A
1       B
2       C
3    AABA
4    BACA
5     NaN
6    CABA
7     DOG
8     CAT
dtype: object

In [4]: s.str.len()
0     1
1     1
2     1
3     4
4     4
5   NaN
6     4
7     3
8     3
dtype: float64

Splitting and Replacing Strings

Methods like split return a Series of lists:

In [5]: s2 = Series(['a_b_c', 'c_d_e', np.nan, 'f_g_h'])

In [6]: s2.str.split('_')
0    [a, b, c]
1    [c, d, e]
2          NaN
3    [f, g, h]
dtype: object

Easy to expand this to return a DataFrame

In [7]: s2.str.split('_').apply(Series)
     0    1    2
0    a    b    c
1    c    d    e
2  NaN  NaN  NaN
3    f    g    h

Elements in the split lists can be accessed using get or [] notation:

In [8]: s2.str.split('_').str.get(1)
0      b
1      d
2    NaN
3      g
dtype: object

In [9]: s2.str.split('_').str[1]
0      b
1      d
2    NaN
3      g
dtype: object

Methods like replace and findall take regular expressions, too:

In [10]: s3 = Series(['A', 'B', 'C', 'Aaba', 'Baca',
   ....:             '', np.nan, 'CABA', 'dog', 'cat'])

In [11]: s3
0       A
1       B
2       C
3    Aaba
4    Baca
6     NaN
7    CABA
8     dog
9     cat
dtype: object

In [12]: s3.str.replace('^.a|dog', 'XX-XX ', case=False)
0           A
1           B
2           C
3    XX-XX ba
4    XX-XX ca
6         NaN
7    XX-XX BA
8      XX-XX 
9     XX-XX t
dtype: object

Some caution must be taken to keep regular expressions in mind! For example, the following code will cause trouble because of the regular expression meaning of $:

# Consider the following badly formatted financial data
In [13]: dollars = Series(['12', '-$10', '$10,000'])

# This does what you'd naively expect:
In [14]: dollars.str.replace('$', '')
0        12
1       -10
2    10,000
dtype: object

# But this doesn't:
In [15]: dollars.str.replace('-$', '-')
0         12
1       -$10
2    $10,000
dtype: object

# We need to escape the special character (for >1 len patterns)
In [16]: dollars.str.replace(r'-\$', '-')
0         12
1        -10
2    $10,000
dtype: object

Indexing with .str

You can use [] notation to directly index by position locations. If you index past the end of the string, the result will be a NaN.

In [17]: s = Series(['A', 'B', 'C', 'Aaba', 'Baca', np.nan,
   ....:             'CABA', 'dog', 'cat'])

In [18]: s.str[0]
0      A
1      B
2      C
3      A
4      B
5    NaN
6      C
7      d
8      c
dtype: object

In [19]: s.str[1]
0    NaN
1    NaN
2    NaN
3      a
4      a
5    NaN
6      A
7      o
8      a
dtype: object

Extracting Substrings

The method extract (introduced in version 0.13) accepts regular expressions with match groups. Extracting a regular expression with one group returns a Series of strings.

In [20]: Series(['a1', 'b2', 'c3']).str.extract('[ab](\d)')
0      1
1      2
2    NaN
dtype: object

Elements that do not match return NaN. Extracting a regular expression with more than one group returns a DataFrame with one column per group.

In [21]: Series(['a1', 'b2', 'c3']).str.extract('([ab])(\d)')
     0    1
0    a    1
1    b    2
2  NaN  NaN

Elements that do not match return a row filled with NaN. Thus, a Series of messy strings can be “converted” into a like-indexed Series or DataFrame of cleaned-up or more useful strings, without necessitating get() to access tuples or re.match objects.

The results dtype always is object, even if no match is found and the result only contains NaN.

Named groups like

In [22]: Series(['a1', 'b2', 'c3']).str.extract('(?P<letter>[ab])(?P<digit>\d)')
  letter digit
0      a     1
1      b     2
2    NaN   NaN

and optional groups like

In [23]: Series(['a1', 'b2', '3']).str.extract('(?P<letter>[ab])?(?P<digit>\d)')
  letter digit
0      a     1
1      b     2
2    NaN     3

can also be used.

Testing for Strings that Match or Contain a Pattern

You can check whether elements contain a pattern:

In [24]: pattern = r'[a-z][0-9]'

In [25]: Series(['1', '2', '3a', '3b', '03c']).str.contains(pattern)
0    False
1    False
2    False
3    False
4    False
dtype: bool

or match a pattern:

In [26]: Series(['1', '2', '3a', '3b', '03c']).str.match(pattern, as_indexer=True)
0    False
1    False
2    False
3    False
4    False
dtype: bool

The distinction between match and contains is strictness: match relies on strict re.match, while contains relies on


In previous versions, match was for extracting groups, returning a not-so-convenient Series of tuples. The new method extract (described in the previous section) is now preferred.

This old, deprecated behavior of match is still the default. As demonstrated above, use the new behavior by setting as_indexer=True. In this mode, match is analogous to contains, returning a boolean Series. The new behavior will become the default behavior in a future release.

Methods like match, contains, startswith, and endswith take
an extra na argument so missing values can be considered True or False:
In [27]: s4 = Series(['A', 'B', 'C', 'Aaba', 'Baca', np.nan, 'CABA', 'dog', 'cat'])

In [28]: s4.str.contains('A', na=False)
0     True
1    False
2    False
3     True
4    False
5    False
6     True
7    False
8    False
dtype: bool

Creating Indicator Variables

You can extract dummy variables from string columns. For example if they are separated by a '|':

In [29]: s = Series(['a', 'a|b', np.nan, 'a|c'])

In [30]: s.str.get_dummies(sep='|')
   a  b  c
0  1  0  0
1  1  1  0
2  0  0  0
3  1  0  1

See also get_dummies().

Method Summary

Method Description
cat() Concatenate strings
split() Split strings on delimiter
get() Index into each element (retrieve i-th element)
join() Join strings in each element of the Series with passed separator
contains() Return boolean array if each string contains pattern/regex
replace() Replace occurrences of pattern/regex with some other string
repeat() Duplicate values (s.str.repeat(3) equivalent to x * 3)
pad() Add whitespace to left, right, or both sides of strings
center() Equivalent to pad(side='both')
wrap() Split long strings into lines with length less than a given width
slice() Slice each string in the Series
slice_replace() Replace slice in each string with passed value
count() Count occurrences of pattern
startswith() Equivalent to str.startswith(pat) for each element
endswith() Equivalent to str.endswith(pat) for each element
findall() Compute list of all occurrences of pattern/regex for each string
match() Call re.match on each element, returning matched groups as list
extract() Call re.match on each element, as match does, but return matched groups as strings for convenience.
len() Compute string lengths
strip() Equivalent to str.strip
rstrip() Equivalent to str.rstrip
lstrip() Equivalent to str.lstrip
lower() Equivalent to str.lower
upper() Equivalent to str.upper