/usr/include/groonga/groonga/tokenizer.h is in libgroonga-dev 6.0.1-1ubuntu1.
This file is owned by root:root, with mode 0o644.
The actual contents of the file can be viewed below.
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 256 257 258 259 260 | /* -*- c-basic-offset: 2 -*- */
/*
Copyright(C) 2012-2016 Brazil
This library is free software; you can redistribute it and/or
modify it under the terms of the GNU Lesser General Public
License version 2.1 as published by the Free Software Foundation.
This library is distributed in the hope that it will be useful,
but WITHOUT ANY WARRANTY; without even the implied warranty of
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
Lesser General Public License for more details.
You should have received a copy of the GNU Lesser General Public
License along with this library; if not, write to the Free Software
Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
*/
#pragma once
#include <groonga/plugin.h>
#include <groonga/token.h>
#ifdef __cplusplus
extern "C" {
#endif /* __cplusplus */
#define GRN_TOKENIZER_TOKENIZED_DELIMITER_UTF8 "\xEF\xBF\xBE"
#define GRN_TOKENIZER_TOKENIZED_DELIMITER_UTF8_LEN 3
/*
grn_tokenizer_charlen() returns the length (#bytes) of the first character
in the string specified by `str_ptr' and `str_length'. If the starting bytes
are invalid as a character, grn_tokenizer_charlen() returns 0. See
grn_encoding in "groonga.h" for more details of `encoding'
Deprecated. Use grn_plugin_charlen() instead.
*/
int grn_tokenizer_charlen(grn_ctx *ctx, const char *str_ptr,
unsigned int str_length, grn_encoding encoding);
/*
grn_tokenizer_isspace() returns the length (#bytes) of the first character
in the string specified by `str_ptr' and `str_length' if it is a space
character. Otherwise, grn_tokenizer_isspace() returns 0.
Deprecated. Use grn_plugin_isspace() instead.
*/
int grn_tokenizer_isspace(grn_ctx *ctx, const char *str_ptr,
unsigned int str_length, grn_encoding encoding);
/*
grn_tokenizer_is_tokenized_delimiter() returns whether is the first
character in the string specified by `str_ptr' and `str_length' the
special tokenized delimiter character or not.
*/
grn_bool grn_tokenizer_is_tokenized_delimiter(grn_ctx *ctx,
const char *str_ptr,
unsigned int str_length,
grn_encoding encoding);
/*
grn_tokenizer_have_tokenized_delimiter() returns whether is there
the special delimiter character in the string specified by `str_ptr'
and `str_length' the special tokenized delimiter character or not.
*/
GRN_PLUGIN_EXPORT grn_bool grn_tokenizer_have_tokenized_delimiter(grn_ctx *ctx,
const char *str_ptr,
unsigned int str_length,
grn_encoding encoding);
/*
grn_tokenizer_query is a structure for storing a query. See the following
functions.
*/
typedef struct _grn_tokenizer_query grn_tokenizer_query;
struct _grn_tokenizer_query {
grn_obj *normalized_query;
char *query_buf;
const char *ptr;
unsigned int length;
grn_encoding encoding;
unsigned int flags;
grn_bool have_tokenized_delimiter;
/* Deprecated since 4.0.8. Use tokenize_mode instead. */
grn_token_mode token_mode;
grn_tokenize_mode tokenize_mode;
};
/*
grn_tokenizer_query_open() parses `args' and returns a new object of
grn_tokenizer_query. The new object stores information of the query.
grn_tokenizer_query_open() normalizes the query if the target table
requires normalization. grn_tokenizer_query_open() returns NULL if
something goes wrong. Note that grn_tokenizer_query_open() must be called
just once in the function that initializes a tokenizer.
See `GRN_STRING_*' flags for `normalize_flags'.
*/
GRN_PLUGIN_EXPORT grn_tokenizer_query *grn_tokenizer_query_open(grn_ctx *ctx,
int num_args, grn_obj **args,
unsigned int normalize_flags);
/*
grn_tokenizer_query_create() is deprecated. Use grn_tokenizer_query_open()
instead.
*/
grn_tokenizer_query *grn_tokenizer_query_create(grn_ctx *ctx,
int num_args, grn_obj **args);
/*
grn_tokenizer_query_close() finalizes an object of grn_tokenizer_query
and then frees memory allocated for that object.
*/
GRN_PLUGIN_EXPORT void grn_tokenizer_query_close(grn_ctx *ctx, grn_tokenizer_query *query);
/*
grn_tokenizer_query_destroy() is deprecated. Use grn_tokenizer_query_close()
instead.
*/
void grn_tokenizer_query_destroy(grn_ctx *ctx, grn_tokenizer_query *query);
/*
grn_tokenizer_token is needed to return tokens. A grn_tokenizer_token object
stores a token to be returned and it must be maintained until a request for
next token or finalization comes.
*/
typedef struct _grn_tokenizer_token grn_tokenizer_token;
struct _grn_tokenizer_token {
grn_obj str;
grn_obj status;
};
/*
grn_tokenizer_token_init() initializes `token'. Note that an initialized
object must be finalized by grn_tokenizer_token_fin().
*/
GRN_PLUGIN_EXPORT void grn_tokenizer_token_init(grn_ctx *ctx, grn_tokenizer_token *token);
/*
grn_tokenizer_token_fin() finalizes `token' that has been initialized by
grn_tokenizer_token_init().
*/
GRN_PLUGIN_EXPORT void grn_tokenizer_token_fin(grn_ctx *ctx, grn_tokenizer_token *token);
/*
* grn_tokenizer_status is a flag set for tokenizer status codes.
* If a document or query contains no tokens, push an empty string with
* GRN_TOKENIZER_TOKEN_LAST as a token.
*
* @deprecated since 4.0.8. Use grn_token_status instead.
*/
typedef grn_token_status grn_tokenizer_status;
/*
* GRN_TOKENIZER_TOKEN_CONTINUE means that the next token is not the last one.
*
* @deprecated since 4.0.8. Use GRN_TOKEN_CONTINUE instead.
*/
#define GRN_TOKENIZER_TOKEN_CONTINUE GRN_TOKEN_CONTINUE
/*
* GRN_TOKENIZER_TOKEN_LAST means that the next token is the last one.
*
* @deprecated since 4.0.8. Use GRN_TOKEN_LAST instead.
*/
#define GRN_TOKENIZER_TOKEN_LAST GRN_TOKEN_LAST
/*
* GRN_TOKENIZER_TOKEN_OVERLAP means that ...
*
* @deprecated since 4.0.8. Use GRN_TOKEN_OVERLAP instead.
*/
#define GRN_TOKENIZER_TOKEN_OVERLAP GRN_TOKEN_OVERLAP
/*
* GRN_TOKENIZER_TOKEN_UNMATURED means that ...
*
* @deprecated since 4.0.8. Use GRN_TOKEN_UNMATURED instead.
*/
#define GRN_TOKENIZER_TOKEN_UNMATURED GRN_TOKEN_UNMATURED
/*
* GRN_TOKENIZER_TOKEN_REACH_END means that ...
*
* @deprecated since 4.0.8. Use GRN_TOKEN_REACH_END instead.
*/
#define GRN_TOKENIZER_TOKEN_REACH_END GRN_TOKEN_REACH_END
/*
* GRN_TOKENIZER_TOKEN_SKIP means that the token is skipped
*
* @deprecated since 4.0.8. Use GRN_TOKEN_SKIP instead.
*/
#define GRN_TOKENIZER_TOKEN_SKIP GRN_TOKEN_SKIP
/*
* GRN_TOKENIZER_TOKEN_SKIP_WITH_POSITION means that the token and postion is skipped
*
* @deprecated since 4.0.8. Use GRN_TOKEN_SKIP_WITH_POSITION instead.
*/
#define GRN_TOKENIZER_TOKEN_SKIP_WITH_POSITION GRN_TOKEN_SKIP_WITH_POSITION
/*
* GRN_TOKENIZER_TOKEN_FORCE_PREIX that the token is used common prefix search
*
* @deprecated since 4.0.8. Use GRN_TOKEN_FORCE_PREIX instead.
*/
#define GRN_TOKENIZER_TOKEN_FORCE_PREFIX GRN_TOKEN_FORCE_PREFIX
/*
* GRN_TOKENIZER_CONTINUE and GRN_TOKENIZER_LAST are deprecated. They
* are just for backward compatibility. Use
* GRN_TOKENIZER_TOKEN_CONTINUE and GRN_TOKENIZER_TOKEN_LAST
* instead.
*/
#define GRN_TOKENIZER_CONTINUE GRN_TOKENIZER_TOKEN_CONTINUE
#define GRN_TOKENIZER_LAST GRN_TOKENIZER_TOKEN_LAST
/*
grn_tokenizer_token_push() pushes the next token into `token'. Note that
grn_tokenizer_token_push() does not make a copy of the given string. This
means that you have to maintain a memory space allocated to the string.
Also note that the grn_tokenizer_token object must be maintained until the
request for the next token or finalization comes. See grn_token_status in
this header for more details of `status'.
*/
GRN_PLUGIN_EXPORT void grn_tokenizer_token_push(grn_ctx *ctx, grn_tokenizer_token *token,
const char *str_ptr, unsigned int str_length,
grn_token_status status);
/*
grn_tokenizer_tokenized_delimiter_next() extracts the next token
from the string specified by `str_ptr' and `str_length' and pushes
the next token into `token'. It returns the string after the next
token. The returned string may be `NULL' when all tokens are
extracted.
*/
GRN_PLUGIN_EXPORT const char *grn_tokenizer_tokenized_delimiter_next(grn_ctx *ctx,
grn_tokenizer_token *token,
const char *str_ptr,
unsigned int str_length,
grn_encoding encoding);
/*
grn_tokenizer_register() registers a plugin to the database which is
associated with `ctx'. `plugin_name_ptr' and `plugin_name_length' specify the
plugin name. Alphabetic letters ('A'-'Z' and 'a'-'z'), digits ('0'-'9') and
an underscore ('_') are capable characters. `init', `next' and `fin' specify
the plugin functions. `init' is called for initializing a tokenizer for a
document or query. `next' is called for extracting tokens one by one. `fin'
is called for finalizing a tokenizer. grn_tokenizer_register() returns
GRN_SUCCESS on success, an error code on failure. See "groonga.h" for more
details of grn_proc_func and grn_user_data, that is used as an argument of
grn_proc_func.
*/
GRN_PLUGIN_EXPORT grn_rc grn_tokenizer_register(grn_ctx *ctx, const char *plugin_name_ptr,
unsigned int plugin_name_length,
grn_proc_func *init, grn_proc_func *next,
grn_proc_func *fin);
#ifdef __cplusplus
} /* extern "C" */
#endif /* __cplusplus */
|