You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
786 lines
31 KiB
786 lines
31 KiB
#!/usr/bin/env python3
|
|
# -*- coding: utf-8 -*-
|
|
#
|
|
# Copyright (C) 2018 The Android Open Source Project
|
|
#
|
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
# you may not use this file except in compliance with the License.
|
|
# You may obtain a copy of the License at
|
|
#
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
#
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
# See the License for the specific language governing permissions and
|
|
# limitations under the License.
|
|
"""A parser for the Minijail policy file."""
|
|
|
|
from __future__ import absolute_import
|
|
from __future__ import division
|
|
from __future__ import print_function
|
|
|
|
import collections
|
|
import itertools
|
|
import os.path
|
|
import re
|
|
|
|
try:
|
|
import bpf
|
|
except ImportError:
|
|
from minijail import bpf
|
|
|
|
|
|
Token = collections.namedtuple(
|
|
'Token', ['type', 'value', 'filename', 'line', 'line_number', 'column'])
|
|
|
|
# A regex that can tokenize a Minijail policy file line.
|
|
_TOKEN_SPECIFICATION = (
|
|
('COMMENT', r'#.*$'),
|
|
('WHITESPACE', r'\s+'),
|
|
('CONTINUATION', r'\\$'),
|
|
('DEFAULT', r'@default\b'),
|
|
('INCLUDE', r'@include\b'),
|
|
('FREQUENCY', r'@frequency\b'),
|
|
('PATH', r'(?:\.)?/\S+'),
|
|
('NUMERIC_CONSTANT', r'-?0[xX][0-9a-fA-F]+|-?0[Oo][0-7]+|-?[0-9]+'),
|
|
('COLON', r':'),
|
|
('SEMICOLON', r';'),
|
|
('COMMA', r','),
|
|
('BITWISE_COMPLEMENT', r'~'),
|
|
('LPAREN', r'\('),
|
|
('RPAREN', r'\)'),
|
|
('LBRACE', r'\{'),
|
|
('RBRACE', r'\}'),
|
|
('RBRACKET', r'\]'),
|
|
('LBRACKET', r'\['),
|
|
('OR', r'\|\|'),
|
|
('AND', r'&&'),
|
|
('BITWISE_OR', r'\|'),
|
|
('OP', r'&|\bin\b|==|!=|<=|<|>=|>'),
|
|
('EQUAL', r'='),
|
|
('ARGUMENT', r'\barg[0-9]+\b'),
|
|
('RETURN', r'\breturn\b'),
|
|
('ACTION',
|
|
r'\ballow\b|\bkill-process\b|\bkill-thread\b|\bkill\b|\btrap\b|'
|
|
r'\btrace\b|\blog\b|\buser-notify\b'
|
|
),
|
|
('IDENTIFIER', r'[a-zA-Z_][a-zA-Z_0-9-@]*'),
|
|
)
|
|
_TOKEN_RE = re.compile('|'.join(
|
|
r'(?P<%s>%s)' % pair for pair in _TOKEN_SPECIFICATION))
|
|
|
|
|
|
class ParseException(Exception):
|
|
"""An exception that is raised when parsing fails."""
|
|
|
|
# pylint: disable=too-many-arguments
|
|
def __init__(self,
|
|
message,
|
|
filename,
|
|
*,
|
|
line='',
|
|
line_number=1,
|
|
token=None):
|
|
if token:
|
|
line = token.line
|
|
line_number = token.line_number
|
|
column = token.column
|
|
length = len(token.value)
|
|
else:
|
|
column = len(line)
|
|
length = 1
|
|
|
|
message = ('%s(%d:%d): %s') % (filename, line_number, column + 1,
|
|
message)
|
|
message += '\n %s' % line
|
|
message += '\n %s%s' % (' ' * column, '^' * length)
|
|
super().__init__(message)
|
|
|
|
|
|
class ParserState:
|
|
"""Stores the state of the Parser to provide better diagnostics."""
|
|
|
|
def __init__(self, filename):
|
|
self._filename = filename
|
|
self._line = ''
|
|
self._line_number = 0
|
|
|
|
@property
|
|
def filename(self):
|
|
"""Return the name of the file being processed."""
|
|
return self._filename
|
|
|
|
@property
|
|
def line(self):
|
|
"""Return the current line being processed."""
|
|
return self._line
|
|
|
|
@property
|
|
def line_number(self):
|
|
"""Return the current line number being processed."""
|
|
return self._line_number
|
|
|
|
def error(self, message, token=None):
|
|
"""Raise a ParserException with the provided message."""
|
|
raise ParseException(
|
|
message,
|
|
self.filename,
|
|
line=self._line,
|
|
line_number=self._line_number,
|
|
token=token)
|
|
|
|
def tokenize(self, lines):
|
|
"""Return a list of tokens for the current line."""
|
|
tokens = []
|
|
|
|
for line_number, line in enumerate(lines):
|
|
self._line_number = line_number + 1
|
|
self._line = line.rstrip('\r\n')
|
|
|
|
last_end = 0
|
|
for token in _TOKEN_RE.finditer(self._line):
|
|
if token.start() != last_end:
|
|
self.error(
|
|
'invalid token',
|
|
token=Token('INVALID',
|
|
self._line[last_end:token.start()],
|
|
self.filename, self._line,
|
|
self._line_number, last_end))
|
|
last_end = token.end()
|
|
|
|
# Omit whitespace and comments now to avoid sprinkling this logic
|
|
# elsewhere.
|
|
if token.lastgroup in ('WHITESPACE', 'COMMENT',
|
|
'CONTINUATION'):
|
|
continue
|
|
tokens.append(
|
|
Token(token.lastgroup, token.group(), self.filename,
|
|
self._line, self._line_number, token.start()))
|
|
if last_end != len(self._line):
|
|
self.error(
|
|
'invalid token',
|
|
token=Token('INVALID', self._line[last_end:],
|
|
self.filename, self._line, self._line_number,
|
|
last_end))
|
|
|
|
if self._line.endswith('\\'):
|
|
# This line is not finished yet.
|
|
continue
|
|
|
|
if tokens:
|
|
# Return a copy of the token list so that the caller can be free
|
|
# to modify it.
|
|
yield tokens[::]
|
|
tokens.clear()
|
|
|
|
|
|
Atom = collections.namedtuple('Atom', ['argument_index', 'op', 'value'])
|
|
"""A single boolean comparison within a filter expression."""
|
|
|
|
Filter = collections.namedtuple('Filter', ['expression', 'action'])
|
|
"""The result of parsing a DNF filter expression, with its action.
|
|
|
|
Since the expression is in Disjunctive Normal Form, it is composed of two levels
|
|
of lists, one for disjunctions and the inner one for conjunctions. The elements
|
|
of the inner list are Atoms.
|
|
"""
|
|
|
|
Syscall = collections.namedtuple('Syscall', ['name', 'number'])
|
|
"""A system call."""
|
|
|
|
ParsedFilterStatement = collections.namedtuple(
|
|
'ParsedFilterStatement', ['syscalls', 'filters', 'token'])
|
|
"""The result of parsing a filter statement.
|
|
|
|
Statements have a list of syscalls, and an associated list of filters that will
|
|
be evaluated sequentially when any of the syscalls is invoked.
|
|
"""
|
|
|
|
FilterStatement = collections.namedtuple('FilterStatement',
|
|
['syscall', 'frequency', 'filters'])
|
|
"""The filter list for a particular syscall.
|
|
|
|
This is a mapping from one syscall to a list of filters that are evaluated
|
|
sequentially. The last filter is always an unconditional action.
|
|
"""
|
|
|
|
ParsedPolicy = collections.namedtuple('ParsedPolicy',
|
|
['default_action', 'filter_statements'])
|
|
"""The result of parsing a minijail .policy file."""
|
|
|
|
|
|
# pylint: disable=too-few-public-methods
|
|
class PolicyParser:
|
|
"""A parser for the Minijail seccomp policy file format."""
|
|
|
|
def __init__(self,
|
|
arch,
|
|
*,
|
|
kill_action,
|
|
include_depth_limit=10,
|
|
override_default_action=None):
|
|
self._parser_states = [ParserState("<memory>")]
|
|
self._kill_action = kill_action
|
|
self._include_depth_limit = include_depth_limit
|
|
self._default_action = self._kill_action
|
|
self._override_default_action = override_default_action
|
|
self._frequency_mapping = collections.defaultdict(int)
|
|
self._arch = arch
|
|
|
|
@property
|
|
def _parser_state(self):
|
|
return self._parser_states[-1]
|
|
|
|
# single-constant = identifier
|
|
# | numeric-constant
|
|
# ;
|
|
def _parse_single_constant(self, token):
|
|
if token.type == 'IDENTIFIER':
|
|
if token.value not in self._arch.constants:
|
|
self._parser_state.error('invalid constant', token=token)
|
|
single_constant = self._arch.constants[token.value]
|
|
elif token.type == 'NUMERIC_CONSTANT':
|
|
try:
|
|
single_constant = int(token.value, base=0)
|
|
except ValueError:
|
|
self._parser_state.error('invalid constant', token=token)
|
|
else:
|
|
self._parser_state.error('invalid constant', token=token)
|
|
if single_constant > self._arch.max_unsigned:
|
|
self._parser_state.error('unsigned overflow', token=token)
|
|
elif single_constant < self._arch.min_signed:
|
|
self._parser_state.error('signed underflow', token=token)
|
|
elif single_constant < 0:
|
|
# This converts the constant to an unsigned representation of the
|
|
# same value, since BPF only uses unsigned values.
|
|
single_constant = self._arch.truncate_word(single_constant)
|
|
return single_constant
|
|
|
|
# constant = [ '~' ] , '(' , value , ')'
|
|
# | [ '~' ] , single-constant
|
|
# ;
|
|
def _parse_constant(self, tokens):
|
|
negate = False
|
|
if tokens[0].type == 'BITWISE_COMPLEMENT':
|
|
negate = True
|
|
tokens.pop(0)
|
|
if not tokens:
|
|
self._parser_state.error('empty complement')
|
|
if tokens[0].type == 'BITWISE_COMPLEMENT':
|
|
self._parser_state.error(
|
|
'invalid double complement', token=tokens[0])
|
|
if tokens[0].type == 'LPAREN':
|
|
last_open_paren = tokens.pop(0)
|
|
single_value = self.parse_value(tokens)
|
|
if not tokens or tokens[0].type != 'RPAREN':
|
|
self._parser_state.error(
|
|
'unclosed parenthesis', token=last_open_paren)
|
|
else:
|
|
single_value = self._parse_single_constant(tokens[0])
|
|
tokens.pop(0)
|
|
if negate:
|
|
single_value = self._arch.truncate_word(~single_value)
|
|
return single_value
|
|
|
|
# value = constant , [ { '|' , constant } ]
|
|
# ;
|
|
def parse_value(self, tokens):
|
|
"""Parse constants separated bitwise OR operator |.
|
|
|
|
Constants can be:
|
|
|
|
- A number that can be parsed with int(..., base=0)
|
|
- A named constant expression.
|
|
- A parenthesized, valid constant expression.
|
|
- A valid constant expression prefixed with the unary bitwise
|
|
complement operator ~.
|
|
- A series of valid constant expressions separated by bitwise
|
|
OR operator |.
|
|
|
|
If there is an error parsing any of the constants, the whole process
|
|
fails.
|
|
"""
|
|
|
|
value = 0
|
|
while tokens:
|
|
value |= self._parse_constant(tokens)
|
|
if not tokens or tokens[0].type != 'BITWISE_OR':
|
|
break
|
|
tokens.pop(0)
|
|
else:
|
|
self._parser_state.error('empty constant')
|
|
return value
|
|
|
|
# atom = argument , op , value
|
|
# ;
|
|
def _parse_atom(self, tokens):
|
|
if not tokens:
|
|
self._parser_state.error('missing argument')
|
|
argument = tokens.pop(0)
|
|
if argument.type != 'ARGUMENT':
|
|
self._parser_state.error('invalid argument', token=argument)
|
|
|
|
if not tokens:
|
|
self._parser_state.error('missing operator')
|
|
operator = tokens.pop(0)
|
|
if operator.type != 'OP':
|
|
self._parser_state.error('invalid operator', token=operator)
|
|
|
|
value = self.parse_value(tokens)
|
|
argument_index = int(argument.value[3:])
|
|
if not (0 <= argument_index < bpf.MAX_SYSCALL_ARGUMENTS):
|
|
self._parser_state.error('invalid argument', token=argument)
|
|
return Atom(argument_index, operator.value, value)
|
|
|
|
# clause = atom , [ { '&&' , atom } ]
|
|
# ;
|
|
def _parse_clause(self, tokens):
|
|
atoms = []
|
|
while tokens:
|
|
atoms.append(self._parse_atom(tokens))
|
|
if not tokens or tokens[0].type != 'AND':
|
|
break
|
|
tokens.pop(0)
|
|
else:
|
|
self._parser_state.error('empty clause')
|
|
return atoms
|
|
|
|
# argument-expression = clause , [ { '||' , clause } ]
|
|
# ;
|
|
def parse_argument_expression(self, tokens):
|
|
"""Parse a argument expression in Disjunctive Normal Form.
|
|
|
|
Since BPF disallows back jumps, we build the basic blocks in reverse
|
|
order so that all the jump targets are known by the time we need to
|
|
reference them.
|
|
"""
|
|
|
|
clauses = []
|
|
while tokens:
|
|
clauses.append(self._parse_clause(tokens))
|
|
if not tokens or tokens[0].type != 'OR':
|
|
break
|
|
tokens.pop(0)
|
|
else:
|
|
self._parser_state.error('empty argument expression')
|
|
return clauses
|
|
|
|
# default-action = 'kill-process'
|
|
# | 'kill-thread'
|
|
# | 'kill'
|
|
# | 'trap'
|
|
# | 'user-notify'
|
|
# ;
|
|
def _parse_default_action(self, tokens):
|
|
if not tokens:
|
|
self._parser_state.error('missing default action')
|
|
action_token = tokens.pop(0)
|
|
if action_token.type != 'ACTION':
|
|
return self._parser_state.error(
|
|
'invalid default action', token=action_token)
|
|
if action_token.value == 'kill-process':
|
|
return bpf.KillProcess()
|
|
if action_token.value == 'kill-thread':
|
|
return bpf.KillThread()
|
|
if action_token.value == 'kill':
|
|
return self._kill_action
|
|
if action_token.value == 'trap':
|
|
return bpf.Trap()
|
|
if action_token.value == 'user-notify':
|
|
return bpf.UserNotify()
|
|
return self._parser_state.error(
|
|
'invalid permissive default action', token=action_token)
|
|
|
|
# action = 'allow' | '1'
|
|
# | 'kill-process'
|
|
# | 'kill-thread'
|
|
# | 'kill'
|
|
# | 'trap'
|
|
# | 'trace'
|
|
# | 'log'
|
|
# | 'user-notify'
|
|
# | 'return' , single-constant
|
|
# ;
|
|
def parse_action(self, tokens):
|
|
if not tokens:
|
|
self._parser_state.error('missing action')
|
|
action_token = tokens.pop(0)
|
|
if action_token.type == 'ACTION':
|
|
if action_token.value == 'allow':
|
|
return bpf.Allow()
|
|
if action_token.value == 'kill':
|
|
return self._kill_action
|
|
if action_token.value == 'kill-process':
|
|
return bpf.KillProcess()
|
|
if action_token.value == 'kill-thread':
|
|
return bpf.KillThread()
|
|
if action_token.value == 'trap':
|
|
return bpf.Trap()
|
|
if action_token.value == 'trace':
|
|
return bpf.Trace()
|
|
if action_token.value == 'user-notify':
|
|
return bpf.UserNotify()
|
|
if action_token.value == 'log':
|
|
return bpf.Log()
|
|
elif action_token.type == 'NUMERIC_CONSTANT':
|
|
constant = self._parse_single_constant(action_token)
|
|
if constant == 1:
|
|
return bpf.Allow()
|
|
elif action_token.type == 'RETURN':
|
|
if not tokens:
|
|
self._parser_state.error('missing return value')
|
|
return bpf.ReturnErrno(self._parse_single_constant(tokens.pop(0)))
|
|
return self._parser_state.error('invalid action', token=action_token)
|
|
|
|
# single-filter = action
|
|
# | argument-expression , [ ';' , action ]
|
|
# ;
|
|
def _parse_single_filter(self, tokens):
|
|
if not tokens:
|
|
self._parser_state.error('missing filter')
|
|
if tokens[0].type == 'ARGUMENT':
|
|
# Only argument expressions can start with an ARGUMENT token.
|
|
argument_expression = self.parse_argument_expression(tokens)
|
|
if tokens and tokens[0].type == 'SEMICOLON':
|
|
tokens.pop(0)
|
|
action = self.parse_action(tokens)
|
|
else:
|
|
action = bpf.Allow()
|
|
return Filter(argument_expression, action)
|
|
else:
|
|
return Filter(None, self.parse_action(tokens))
|
|
|
|
# filter = '{' , single-filter , [ { ',' , single-filter } ] , '}'
|
|
# | single-filter
|
|
# ;
|
|
def parse_filter(self, tokens):
|
|
"""Parse a filter and return a list of Filter objects."""
|
|
if not tokens:
|
|
self._parser_state.error('missing filter')
|
|
filters = []
|
|
if tokens[0].type == 'LBRACE':
|
|
opening_brace = tokens.pop(0)
|
|
while tokens:
|
|
filters.append(self._parse_single_filter(tokens))
|
|
if not tokens or tokens[0].type != 'COMMA':
|
|
break
|
|
tokens.pop(0)
|
|
if not tokens or tokens[0].type != 'RBRACE':
|
|
self._parser_state.error('unclosed brace', token=opening_brace)
|
|
tokens.pop(0)
|
|
else:
|
|
filters.append(self._parse_single_filter(tokens))
|
|
return filters
|
|
|
|
# key-value-pair = identifier , '=', identifier , [ { ',' , identifier } ]
|
|
# ;
|
|
def _parse_key_value_pair(self, tokens):
|
|
if not tokens:
|
|
self._parser_state.error('missing key')
|
|
key = tokens.pop(0)
|
|
if key.type != 'IDENTIFIER':
|
|
self._parser_state.error('invalid key', token=key)
|
|
if not tokens:
|
|
self._parser_state.error('missing equal')
|
|
if tokens[0].type != 'EQUAL':
|
|
self._parser_state.error('invalid equal', token=tokens[0])
|
|
tokens.pop(0)
|
|
value_list = []
|
|
while tokens:
|
|
value = tokens.pop(0)
|
|
if value.type != 'IDENTIFIER':
|
|
self._parser_state.error('invalid value', token=value)
|
|
value_list.append(value.value)
|
|
if not tokens or tokens[0].type != 'COMMA':
|
|
break
|
|
tokens.pop(0)
|
|
else:
|
|
self._parser_state.error('empty value')
|
|
return (key.value, value_list)
|
|
|
|
# metadata = '[' , key-value-pair , [ { ';' , key-value-pair } ] , ']'
|
|
# ;
|
|
def _parse_metadata(self, tokens):
|
|
if not tokens:
|
|
self._parser_state.error('missing opening bracket')
|
|
opening_bracket = tokens.pop(0)
|
|
if opening_bracket.type != 'LBRACKET':
|
|
self._parser_state.error(
|
|
'invalid opening bracket', token=opening_bracket)
|
|
metadata = {}
|
|
while tokens:
|
|
first_token = tokens[0]
|
|
key, value = self._parse_key_value_pair(tokens)
|
|
if key in metadata:
|
|
self._parser_state.error(
|
|
'duplicate metadata key: "%s"' % key, token=first_token)
|
|
metadata[key] = value
|
|
if not tokens or tokens[0].type != 'SEMICOLON':
|
|
break
|
|
tokens.pop(0)
|
|
if not tokens or tokens[0].type != 'RBRACKET':
|
|
self._parser_state.error('unclosed bracket', token=opening_bracket)
|
|
tokens.pop(0)
|
|
return metadata
|
|
|
|
# syscall-descriptor = syscall-name , [ metadata ]
|
|
# | syscall-group-name , [ metadata ]
|
|
# ;
|
|
def _parse_syscall_descriptor(self, tokens):
|
|
if not tokens:
|
|
self._parser_state.error('missing syscall descriptor')
|
|
syscall_descriptor = tokens.pop(0)
|
|
# `kill` as a syscall name is a special case since kill is also a valid
|
|
# action and actions have precendence over identifiers.
|
|
if (syscall_descriptor.type != 'IDENTIFIER' and
|
|
syscall_descriptor.value != 'kill'):
|
|
self._parser_state.error(
|
|
'invalid syscall descriptor', token=syscall_descriptor)
|
|
if tokens and tokens[0].type == 'LBRACKET':
|
|
metadata = self._parse_metadata(tokens)
|
|
if 'arch' in metadata and self._arch.arch_name not in metadata['arch']:
|
|
return ()
|
|
if '@' in syscall_descriptor.value:
|
|
# This is a syscall group.
|
|
subtokens = syscall_descriptor.value.split('@')
|
|
if len(subtokens) != 2:
|
|
self._parser_state.error(
|
|
'invalid syscall group name', token=syscall_descriptor)
|
|
syscall_group_name, syscall_namespace_name = subtokens
|
|
if syscall_namespace_name not in self._arch.syscall_groups:
|
|
self._parser_state.error(
|
|
'nonexistent syscall group namespace',
|
|
token=syscall_descriptor)
|
|
syscall_namespace = self._arch.syscall_groups[
|
|
syscall_namespace_name]
|
|
if syscall_group_name not in syscall_namespace:
|
|
self._parser_state.error(
|
|
'nonexistent syscall group', token=syscall_descriptor)
|
|
return (Syscall(name, self._arch.syscalls[name])
|
|
for name in syscall_namespace[syscall_group_name])
|
|
if syscall_descriptor.value not in self._arch.syscalls:
|
|
self._parser_state.error(
|
|
'nonexistent syscall', token=syscall_descriptor)
|
|
return (Syscall(syscall_descriptor.value,
|
|
self._arch.syscalls[syscall_descriptor.value]), )
|
|
|
|
# filter-statement = '{' , syscall-descriptor , [ { ',', syscall-descriptor } ] , '}' ,
|
|
# ':' , filter
|
|
# | syscall-descriptor , ':' , filter
|
|
# ;
|
|
def parse_filter_statement(self, tokens):
|
|
"""Parse a filter statement and return a ParsedFilterStatement."""
|
|
if not tokens:
|
|
self._parser_state.error('empty filter statement')
|
|
syscall_descriptors = []
|
|
if tokens[0].type == 'LBRACE':
|
|
opening_brace = tokens.pop(0)
|
|
while tokens:
|
|
syscall_descriptors.extend(
|
|
self._parse_syscall_descriptor(tokens))
|
|
if not tokens or tokens[0].type != 'COMMA':
|
|
break
|
|
tokens.pop(0)
|
|
if not tokens or tokens[0].type != 'RBRACE':
|
|
self._parser_state.error('unclosed brace', token=opening_brace)
|
|
tokens.pop(0)
|
|
else:
|
|
syscall_descriptors.extend(self._parse_syscall_descriptor(tokens))
|
|
if not tokens:
|
|
self._parser_state.error('missing colon')
|
|
if tokens[0].type != 'COLON':
|
|
self._parser_state.error('invalid colon', token=tokens[0])
|
|
# Given that there can be multiple syscalls and filters in a single
|
|
# filter statement, use the colon token as the anchor for error location
|
|
# purposes.
|
|
colon_token = tokens.pop(0)
|
|
parsed_filter = self.parse_filter(tokens)
|
|
if not syscall_descriptors:
|
|
return None
|
|
return ParsedFilterStatement(
|
|
tuple(syscall_descriptors), parsed_filter, colon_token)
|
|
|
|
# include-statement = '@include' , posix-path
|
|
# ;
|
|
def _parse_include_statement(self, tokens):
|
|
if not tokens:
|
|
self._parser_state.error('empty filter statement')
|
|
if tokens[0].type != 'INCLUDE':
|
|
self._parser_state.error('invalid include', token=tokens[0])
|
|
tokens.pop(0)
|
|
if not tokens:
|
|
self._parser_state.error('empty include path')
|
|
include_path = tokens.pop(0)
|
|
if include_path.type != 'PATH':
|
|
self._parser_state.error(
|
|
'invalid include path', token=include_path)
|
|
if len(self._parser_states) == self._include_depth_limit:
|
|
self._parser_state.error('@include statement nested too deep')
|
|
include_filename = os.path.normpath(
|
|
os.path.join(
|
|
os.path.dirname(self._parser_state.filename),
|
|
include_path.value))
|
|
if not os.path.isfile(include_filename):
|
|
self._parser_state.error(
|
|
'Could not @include %s' % include_filename, token=include_path)
|
|
return self._parse_policy_file(include_filename)
|
|
|
|
def _parse_frequency_file(self, filename):
|
|
self._parser_states.append(ParserState(filename))
|
|
try:
|
|
frequency_mapping = collections.defaultdict(int)
|
|
with open(filename) as frequency_file:
|
|
for tokens in self._parser_state.tokenize(frequency_file):
|
|
syscall_numbers = self._parse_syscall_descriptor(tokens)
|
|
if not tokens:
|
|
self._parser_state.error('missing colon')
|
|
if tokens[0].type != 'COLON':
|
|
self._parser_state.error(
|
|
'invalid colon', token=tokens[0])
|
|
tokens.pop(0)
|
|
|
|
if not tokens:
|
|
self._parser_state.error('missing number')
|
|
number = tokens.pop(0)
|
|
if number.type != 'NUMERIC_CONSTANT':
|
|
self._parser_state.error(
|
|
'invalid number', token=number)
|
|
number_value = int(number.value, base=0)
|
|
if number_value < 0:
|
|
self._parser_state.error(
|
|
'invalid number', token=number)
|
|
|
|
for syscall_number in syscall_numbers:
|
|
frequency_mapping[syscall_number] += number_value
|
|
return frequency_mapping
|
|
finally:
|
|
self._parser_states.pop()
|
|
|
|
# frequency-statement = '@frequency' , posix-path
|
|
# ;
|
|
def _parse_frequency_statement(self, tokens):
|
|
if not tokens:
|
|
self._parser_state.error('empty frequency statement')
|
|
if tokens[0].type != 'FREQUENCY':
|
|
self._parser_state.error('invalid frequency', token=tokens[0])
|
|
tokens.pop(0)
|
|
if not tokens:
|
|
self._parser_state.error('empty frequency path')
|
|
frequency_path = tokens.pop(0)
|
|
if frequency_path.type != 'PATH':
|
|
self._parser_state.error(
|
|
'invalid frequency path', token=frequency_path)
|
|
frequency_filename = os.path.normpath(
|
|
os.path.join(
|
|
os.path.dirname(self._parser_state.filename),
|
|
frequency_path.value))
|
|
if not os.path.isfile(frequency_filename):
|
|
self._parser_state.error(
|
|
'Could not open frequency file %s' % frequency_filename,
|
|
token=frequency_path)
|
|
return self._parse_frequency_file(frequency_filename)
|
|
|
|
# default-statement = '@default' , default-action
|
|
# ;
|
|
def _parse_default_statement(self, tokens):
|
|
if not tokens:
|
|
self._parser_state.error('empty default statement')
|
|
if tokens[0].type != 'DEFAULT':
|
|
self._parser_state.error('invalid default', token=tokens[0])
|
|
tokens.pop(0)
|
|
if not tokens:
|
|
self._parser_state.error('empty action')
|
|
return self._parse_default_action(tokens)
|
|
|
|
def _parse_policy_file(self, filename):
|
|
self._parser_states.append(ParserState(filename))
|
|
try:
|
|
statements = []
|
|
with open(filename) as policy_file:
|
|
for tokens in self._parser_state.tokenize(policy_file):
|
|
if tokens[0].type == 'INCLUDE':
|
|
statements.extend(
|
|
self._parse_include_statement(tokens))
|
|
elif tokens[0].type == 'FREQUENCY':
|
|
for syscall_number, frequency in self._parse_frequency_statement(
|
|
tokens).items():
|
|
self._frequency_mapping[
|
|
syscall_number] += frequency
|
|
elif tokens[0].type == 'DEFAULT':
|
|
self._default_action = self._parse_default_statement(
|
|
tokens)
|
|
else:
|
|
statement = self.parse_filter_statement(tokens)
|
|
if statement is None:
|
|
# If all the syscalls in the statement are for
|
|
# another arch, skip the whole statement.
|
|
continue
|
|
statements.append(statement)
|
|
|
|
if tokens:
|
|
self._parser_state.error(
|
|
'extra tokens', token=tokens[0])
|
|
return statements
|
|
finally:
|
|
self._parser_states.pop()
|
|
|
|
def parse_file(self, filename):
|
|
"""Parse a file and return the list of FilterStatements."""
|
|
self._frequency_mapping = collections.defaultdict(int)
|
|
try:
|
|
statements = [x for x in self._parse_policy_file(filename)]
|
|
except RecursionError:
|
|
raise ParseException(
|
|
'recursion limit exceeded',
|
|
filename,
|
|
line=self._parser_states[-1].line)
|
|
|
|
# Collapse statements into a single syscall-to-filter-list, remembering
|
|
# the token for each filter for better diagnostics.
|
|
syscall_filter_mapping = {}
|
|
syscall_filter_definitions = {}
|
|
filter_statements = []
|
|
for syscalls, filters, token in statements:
|
|
for syscall in syscalls:
|
|
if syscall not in syscall_filter_mapping:
|
|
filter_statements.append(
|
|
FilterStatement(
|
|
syscall, self._frequency_mapping.get(syscall, 1),
|
|
[]))
|
|
syscall_filter_mapping[syscall] = filter_statements[-1]
|
|
syscall_filter_definitions[syscall] = []
|
|
for filt in filters:
|
|
syscall_filter_mapping[syscall].filters.append(filt)
|
|
syscall_filter_definitions[syscall].append(token)
|
|
default_action = self._override_default_action or self._default_action
|
|
for filter_statement in filter_statements:
|
|
unconditional_actions_suffix = list(
|
|
itertools.dropwhile(lambda filt: filt.expression is not None,
|
|
filter_statement.filters))
|
|
if len(unconditional_actions_suffix) == 1:
|
|
# The last filter already has an unconditional action, no need
|
|
# to add another one.
|
|
continue
|
|
if len(unconditional_actions_suffix) > 1:
|
|
previous_definition_token = syscall_filter_definitions[
|
|
filter_statement.syscall][
|
|
-len(unconditional_actions_suffix)]
|
|
current_definition_token = syscall_filter_definitions[
|
|
filter_statement.syscall][
|
|
-len(unconditional_actions_suffix) + 1]
|
|
raise ParseException(
|
|
('Syscall %s (number %d) already had '
|
|
'an unconditional action applied') %
|
|
(filter_statement.syscall.name,
|
|
filter_statement.syscall.number),
|
|
filename=current_definition_token.filename,
|
|
token=current_definition_token) from ParseException(
|
|
'Previous definition',
|
|
filename=previous_definition_token.filename,
|
|
token=previous_definition_token)
|
|
assert not unconditional_actions_suffix
|
|
filter_statement.filters.append(
|
|
Filter(expression=None, action=default_action))
|
|
return ParsedPolicy(default_action, filter_statements)
|