1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
|
/*
Copyright (c) 2012-2018 Red Hat, Inc. <http://www.redhat.com>
This file is part of GlusterFS.
This file is licensed to you under your choice of the GNU Lesser
General Public License, version 3 or any later version (LGPLv3 or
later), or the GNU General Public License, version 2 (GPLv2), in all
cases as published by the Free Software Foundation.
*/
#ifndef _GLFS_INTERNAL_H
#define _GLFS_INTERNAL_H
#include <glusterfs/xlator.h>
#include <glusterfs/glusterfs.h>
#include <glusterfs/upcall-utils.h>
#include "glfs-handles.h"
#include <glusterfs/refcount.h>
#define GLFS_SYMLINK_MAX_FOLLOW 2048
#define DEFAULT_REVAL_COUNT 1
/*
* According to pthread mutex and conditional variable ( cond,
* child_down_count, upcall mutex and mutex) initialization of struct glfs
* members, below GLFS_INIT_* flags are set in 'pthread_flags' member of struct
* glfs. The flags are set from glfs_init() and glfs_new_from_ctx() functions
* as part of fs inititialization.
*
* These flag bits are validated in glfs_fini() to destroy all or partially
* initialized mutex and conditional variables of glfs object.
* If you introduce new pthread mutex or conditional variable in glfs object,
* please make sure you have a flag bit intorduced here for proper cleanup
* in glfs_fini().
*
*/
#define PTHREAD_MUTEX_INIT(mutex, attr, flags, mask, label) \
do { \
int __ret = -1; \
__ret = pthread_mutex_init(mutex, attr); \
if (__ret == 0) \
flags |= mask; \
else \
goto label; \
} while (0)
#define PTHREAD_MUTEX_DESTROY(mutex, flags, mask) \
do { \
if (flags & mask) \
(void)pthread_mutex_destroy(mutex); \
} while (0)
#define PTHREAD_COND_INIT(cond, attr, flags, mask, label) \
do { \
int __ret = -1; \
__ret = pthread_cond_init(cond, attr); \
if (__ret == 0) \
flags |= mask; \
else \
goto label; \
} while (0)
#define PTHREAD_COND_DESTROY(cond, flags, mask) \
do { \
if (flags & mask) \
(void)pthread_cond_destroy(cond); \
} while (0)
#define GLFS_INIT_MUTEX 0x00000001 /* pthread_mutex_flag */
#define GLFS_INIT_COND 0x00000002 /* pthread_cond_flag */
#define GLFS_INIT_COND_CHILD 0x00000004 /* pthread_cond_child_down_flag */
#define GLFS_INIT_MUTEX_UPCALL 0x00000008 /* pthread_mutex_upcall_flag */
#ifndef GF_DARWIN_HOST_OS
#ifndef GFAPI_PUBLIC
#define GFAPI_PUBLIC(sym, ver) /**/
#endif
#ifndef GFAPI_PRIVATE
#define GFAPI_PRIVATE(sym, ver) /**/
#endif
#define GFAPI_SYMVER_PUBLIC_DEFAULT(fn, ver) \
asm(".symver pub_" STR(fn) ", " STR(fn) "@@GFAPI_" STR(ver))
#define GFAPI_SYMVER_PRIVATE_DEFAULT(fn, ver) \
asm(".symver priv_" STR(fn) ", " STR(fn) "@@GFAPI_PRIVATE_" STR(ver))
#define GFAPI_SYMVER_PUBLIC(fn1, fn2, ver) \
asm(".symver pub_" STR(fn1) ", " STR(fn2) "@GFAPI_" STR(ver))
#define GFAPI_SYMVER_PRIVATE(fn1, fn2, ver) \
asm(".symver priv_" STR(fn1) ", " STR(fn2) "@GFAPI_PRIVATE_" STR(ver))
#define STR(str) #str
#else
#ifndef GFAPI_PUBLIC
#define GFAPI_PUBLIC(sym, ver) __asm("_" __STRING(sym) "$GFAPI_" __STRING(ver))
#endif
#ifndef GFAPI_PRIVATE
#define GFAPI_PRIVATE(sym, ver) \
__asm("_" __STRING(sym) "$GFAPI_PRIVATE_" __STRING(ver))
#endif
#define GFAPI_SYMVER_PUBLIC_DEFAULT(fn, dotver) /**/
#define GFAPI_SYMVER_PRIVATE_DEFAULT(fn, dotver) /**/
#define GFAPI_SYMVER_PUBLIC(fn1, fn2, dotver) /**/
#define GFAPI_SYMVER_PRIVATE(fn1, fn2, dotver) /**/
#endif
#define ESTALE_RETRY(ret, errno, reval, loc, label) \
do { \
if (ret == -1 && errno == ESTALE) { \
if (reval < DEFAULT_REVAL_COUNT) { \
reval++; \
loc_wipe(loc); \
goto label; \
} \
} \
} while (0)
#define GLFS_LOC_FILL_INODE(oinode, loc, label) \
do { \
loc.inode = inode_ref(oinode); \
gf_uuid_copy(loc.gfid, oinode->gfid); \
ret = glfs_loc_touchup(&loc); \
if (ret != 0) { \
errno = EINVAL; \
goto label; \
} \
} while (0)
#define GLFS_LOC_FILL_PINODE(pinode, loc, ret, errno, label, path) \
do { \
loc.inode = inode_new(pinode->table); \
if (!loc.inode) { \
ret = -1; \
errno = ENOMEM; \
goto label; \
} \
loc.parent = inode_ref(pinode); \
loc.name = path; \
ret = glfs_loc_touchup(&loc); \
if (ret != 0) { \
errno = EINVAL; \
goto label; \
} \
} while (0)
struct glfs;
struct _upcall_entry {
struct list_head upcall_list;
struct gf_upcall upcall_data;
};
typedef struct _upcall_entry upcall_entry;
typedef int (*glfs_init_cbk)(struct glfs *fs, int ret);
struct glfs {
char *volname;
uuid_t vol_uuid;
glusterfs_ctx_t *ctx;
pthread_t poller;
glfs_init_cbk init_cbk;
pthread_mutex_t mutex;
pthread_cond_t cond;
pthread_cond_t child_down_cond; /* for broadcasting CHILD_DOWN */
int init;
int ret;
int err;
xlator_t *active_subvol; /* active graph */
xlator_t *mip_subvol; /* graph for which migration is in
* progress */
xlator_t *next_subvol; /* Any new graph is put to
* next_subvol, the graph in
* next_subvol can either be moved
* to mip_subvol (if any IO picks it
* up for migration), or be
* destroyed (if there is a new
* graph, and this was never picked
* for migration) */
xlator_t *old_subvol;
char *oldvolfile;
ssize_t oldvollen;
inode_t *cwd;
uint32_t dev_id; /* Used to fill st_dev in struct stat */
struct list_head openfds;
gf_boolean_t migration_in_progress;
gf_boolean_t cache_upcalls; /* add upcalls to the upcall_list? */
struct list_head upcall_list;
pthread_mutex_t upcall_list_mutex; /* mutex for upcall entry list */
uint32_t pin_refcnt;
uint32_t pthread_flags; /* GLFS_INIT_* # defines set this flag */
uint32_t upcall_events; /* Mask of upcall events application
* is interested in */
glfs_upcall_cbk up_cbk; /* upcall cbk function to be registered */
void *up_data; /* Opaque data provided by application
* during upcall registration */
};
/* This enum is used to maintain the state of glfd. In case of async fops
* fd might be closed before the actual fop is complete. Therefore we need
* to track whether the fd is closed or not, instead actually closing it.*/
enum glfs_fd_state { GLFD_INIT, GLFD_OPEN, GLFD_CLOSE };
struct glfs_fd {
struct list_head openfds;
struct list_head list;
GF_REF_DECL;
struct glfs *fs;
enum glfs_fd_state state;
off_t offset;
fd_t *fd; /* Currently guared by @fs->mutex. TODO: per-glfd lock */
struct list_head entries;
gf_dirent_t *next;
struct dirent *readdirbuf;
gf_lkowner_t lk_owner;
glfs_leaseid_t lease_id; /* Stores lease_id of client in glfd */
gf_lock_t lock; /* lock taken before updating fd state */
glfs_recall_cbk cbk;
void *cookie;
};
/* glfs object handle introduced for the alternate gfapi implementation based
on glfs handles/gfid/inode
*/
struct glfs_object {
inode_t *inode;
uuid_t gfid;
};
struct glfs_upcall {
struct glfs *fs; /* glfs object */
enum glfs_upcall_reason reason; /* Upcall event type */
void *event; /* changes based in the event type */
void (*free_event)(void *); /* free event after the usage */
};
struct glfs_upcall_inode {
struct glfs_object *object; /* Object which need to be acted upon */
int flags; /* Cache UPDATE/INVALIDATE flags */
struct stat buf; /* Latest stat of this entry */
unsigned int expire_time_attr; /* the amount of time for which
* the application need to cache
* this entry */
struct glfs_object *p_object; /* parent Object to be updated */
struct stat p_buf; /* Latest stat of parent dir handle */
struct glfs_object *oldp_object; /* Old parent Object to be updated */
struct stat oldp_buf; /* Latest stat of old parent dir handle */
};
struct glfs_upcall_lease {
struct glfs_object *object; /* Object which need to be acted upon */
uint32_t lease_type; /* Lease type to which client can downgrade to*/
};
struct glfs_upcall_lease_fd {
uint32_t lease_type; /* Lease type to which client can downgrade to*/
void *fd_cookie; /* Object which need to be acted upon */
};
struct glfs_xreaddirp_stat {
struct stat
st; /* Stat for that dirent - corresponds to GFAPI_XREADDIRP_STAT */
struct glfs_object *object; /* handled for GFAPI_XREADDIRP_HANDLE */
uint32_t flags_handled; /* final set of flags successfulyy handled */
};
#define DEFAULT_EVENT_POOL_SIZE 16384
#define GF_MEMPOOL_COUNT_OF_DICT_T 4096
#define GF_MEMPOOL_COUNT_OF_DATA_T (GF_MEMPOOL_COUNT_OF_DICT_T * 4)
#define GF_MEMPOOL_COUNT_OF_DATA_PAIR_T (GF_MEMPOOL_COUNT_OF_DICT_T * 4)
#define GF_MEMPOOL_COUNT_OF_LRU_BUF_T 256
typedef void(glfs_mem_release_t)(void *ptr);
struct glfs_mem_header {
uint32_t magic;
size_t nmemb;
size_t size;
glfs_mem_release_t *release;
};
#define GLFS_MEM_HEADER_SIZE (sizeof(struct glfs_mem_header))
#define GLFS_MEM_HEADER_MAGIC 0x20170830
static inline void *
__glfs_calloc(size_t nmemb, size_t size, glfs_mem_release_t release,
uint32_t type, const char *typestr)
{
struct glfs_mem_header *header = NULL;
header = __gf_calloc(nmemb, (size + GLFS_MEM_HEADER_SIZE), type, typestr);
if (!header)
return NULL;
header->magic = GLFS_MEM_HEADER_MAGIC;
header->nmemb = nmemb;
header->size = size;
header->release = release;
return header + 1;
}
static inline void *
__glfs_malloc(size_t size, glfs_mem_release_t release, uint32_t type,
const char *typestr)
{
struct glfs_mem_header *header = NULL;
header = __gf_malloc((size + GLFS_MEM_HEADER_SIZE), type, typestr);
if (!header)
return NULL;
header->magic = GLFS_MEM_HEADER_MAGIC;
header->nmemb = 1;
header->size = size;
header->release = release;
return header + 1;
}
static inline void *
__glfs_realloc(void *ptr, size_t size)
{
struct glfs_mem_header *old_header = NULL;
struct glfs_mem_header *new_header = NULL;
struct glfs_mem_header tmp_header;
void *new_ptr = NULL;
GF_ASSERT(NULL != ptr);
old_header = (struct glfs_mem_header *)(ptr - GLFS_MEM_HEADER_SIZE);
GF_ASSERT(old_header->magic == GLFS_MEM_HEADER_MAGIC);
tmp_header = *old_header;
new_ptr = __gf_realloc(old_header, (size + GLFS_MEM_HEADER_SIZE));
if (!new_ptr)
return NULL;
new_header = (struct glfs_mem_header *)new_ptr;
*new_header = tmp_header;
new_header->size = size;
return new_header + 1;
}
static inline void
__glfs_free(void *free_ptr)
{
struct glfs_mem_header *header = NULL;
void *release_ptr = NULL;
int i = 0;
if (!free_ptr)
return;
header = (struct glfs_mem_header *)(free_ptr - GLFS_MEM_HEADER_SIZE);
GF_ASSERT(header->magic == GLFS_MEM_HEADER_MAGIC);
if (header->release) {
release_ptr = free_ptr;
for (i = 0; i < header->nmemb; i++) {
header->release(release_ptr);
release_ptr += header->size;
}
}
__gf_free(header);
}
#define GLFS_CALLOC(nmemb, size, release, type) \
__glfs_calloc(nmemb, size, release, type, #type)
#define GLFS_MALLOC(size, release, type) \
__glfs_malloc(size, release, type, #type)
#define GLFS_REALLOC(ptr, size) __glfs_realloc(ptr, size)
#define GLFS_FREE(free_ptr) __glfs_free(free_ptr)
int
glfs_mgmt_init(struct glfs *fs);
void
glfs_init_done(struct glfs *fs, int ret) GFAPI_PRIVATE(glfs_init_done, 3.4.0);
int
glfs_process_volfp(struct glfs *fs, FILE *fp);
int
glfs_resolve(struct glfs *fs, xlator_t *subvol, const char *path, loc_t *loc,
struct iatt *iatt, int reval) GFAPI_PRIVATE(glfs_resolve, 3.7.0);
int
glfs_lresolve(struct glfs *fs, xlator_t *subvol, const char *path, loc_t *loc,
struct iatt *iatt, int reval);
fd_t *
glfs_resolve_fd(struct glfs *fs, xlator_t *subvol, struct glfs_fd *glfd);
fd_t *
__glfs_migrate_fd(struct glfs *fs, xlator_t *subvol, struct glfs_fd *glfd);
int
glfs_first_lookup(xlator_t *subvol);
void
glfs_process_upcall_event(struct glfs *fs, void *data)
GFAPI_PRIVATE(glfs_process_upcall_event, 3.7.0);
#define __GLFS_ENTRY_VALIDATE_FS(fs, label) \
do { \
if (!fs) { \
errno = EINVAL; \
goto label; \
} \
old_THIS = THIS; \
THIS = fs->ctx->master; \
} while (0)
#define __GLFS_EXIT_FS \
do { \
THIS = old_THIS; \
} while (0)
#define __GLFS_ENTRY_VALIDATE_FD(glfd, label) \
do { \
if (!glfd || !glfd->fd || !glfd->fd->inode || \
glfd->state != GLFD_OPEN) { \
errno = EBADF; \
goto label; \
} \
old_THIS = THIS; \
THIS = glfd->fd->inode->table->xl->ctx->master; \
} while (0)
/*
By default all lock attempts from user context must
use glfs_lock() and glfs_unlock(). This allows
for a safe implementation of graph migration where
we can give up the mutex during syncop calls so
that bottom up calls (particularly CHILD_UP notify)
can do a mutex_lock() on @glfs without deadlocking
the filesystem.
All the fops should wait for graph migration to finish
before starting the fops. Therefore these functions should
call glfs_lock with wait_for_migration as true. But waiting
for migration to finish in call-back path can result thread
dead-locks. The reason for this is we only have finite
number of epoll threads. so if we wait on epoll threads
there will not be any thread left to handle outstanding
rpc replies.
*/
static inline int
glfs_lock(struct glfs *fs, gf_boolean_t wait_for_migration)
{
pthread_mutex_lock(&fs->mutex);
while (!fs->init)
pthread_cond_wait(&fs->cond, &fs->mutex);
while (wait_for_migration && fs->migration_in_progress)
pthread_cond_wait(&fs->cond, &fs->mutex);
return 0;
}
static inline void
glfs_unlock(struct glfs *fs)
{
pthread_mutex_unlock(&fs->mutex);
}
struct glfs_fd *
glfs_fd_new(struct glfs *fs);
void
glfs_fd_bind(struct glfs_fd *glfd);
void
glfd_set_state_bind(struct glfs_fd *glfd);
xlator_t *
glfs_active_subvol(struct glfs *fs) GFAPI_PRIVATE(glfs_active_subvol, 3.4.0);
xlator_t *
__glfs_active_subvol(struct glfs *fs);
void
glfs_subvol_done(struct glfs *fs, xlator_t *subvol)
GFAPI_PRIVATE(glfs_subvol_done, 3.4.0);
inode_t *
glfs_refresh_inode(xlator_t *subvol, inode_t *inode);
inode_t *
glfs_cwd_get(struct glfs *fs);
int
glfs_cwd_set(struct glfs *fs, inode_t *inode);
inode_t *
glfs_resolve_inode(struct glfs *fs, xlator_t *subvol,
struct glfs_object *object);
int
glfs_create_object(loc_t *loc, struct glfs_object **retobject);
int
__glfs_cwd_set(struct glfs *fs, inode_t *inode);
int
glfs_resolve_base(struct glfs *fs, xlator_t *subvol, inode_t *inode,
struct iatt *iatt);
int
glfs_resolve_at(struct glfs *fs, xlator_t *subvol, inode_t *at,
const char *origpath, loc_t *loc, struct iatt *iatt, int follow,
int reval) GFAPI_PRIVATE(glfs_resolve_at, 3.4.0);
int
glfs_loc_touchup(loc_t *loc) GFAPI_PRIVATE(glfs_loc_touchup, 3.4.0);
void
glfs_iatt_to_stat(struct glfs *fs, struct iatt *iatt, struct stat *stat);
int
glfs_loc_link(loc_t *loc, struct iatt *iatt);
int
glfs_loc_unlink(loc_t *loc);
int
glfs_getxattr_process(void *value, size_t size, dict_t *xattr,
const char *name);
/* Sends RPC call to glusterd to fetch required volume info */
int
glfs_get_volume_info(struct glfs *fs);
/*
SYNOPSIS
glfs_new_from_ctx: Creates a virtual mount object by taking a
glusterfs_ctx_t object.
DESCRIPTION
glfs_new_from_ctx() is not same as glfs_new(). It takes the
glusterfs_ctx_t object instead of creating one by glusterfs_ctx_new().
Again the usage is restricted to NFS MOUNT over UDP i.e. in
glfs_resolve_at() which would take fs object as input but never use
(purpose is not to change the ABI of glfs_resolve_at()).
PARAMETERS
@ctx: glusterfs_ctx_t object
RETURN VALUES
fs : Pointer to the newly created glfs_t object.
NULL : Otherwise.
*/
struct glfs *
glfs_new_from_ctx(glusterfs_ctx_t *ctx) GFAPI_PRIVATE(glfs_new_from_ctx, 3.7.0);
/*
SYNOPSIS
glfs_free_from_ctx: Free up the memory occupied by glfs_t object
created by glfs_new_from_ctx().
DESCRIPTION
The glfs_t object allocated by glfs_new_from_ctx() must be released
by the caller using this routine. The usage can be found
at glfs_fini() or NFS, MOUNT over UDP i.e.
__mnt3udp_get_export_subdir_inode ()
=> glfs_resolve_at().
PARAMETERS
@fs: The glfs_t object to be deallocated.
RETURN VALUES
void
*/
void
glfs_free_from_ctx(struct glfs *fs) GFAPI_PRIVATE(glfs_free_from_ctx, 3.7.0);
int
glfs_recall_lease_fd(struct glfs *fs, struct gf_upcall *up_data);
int
glfs_get_upcall_cache_invalidation(struct gf_upcall *to_up_data,
struct gf_upcall *from_up_data);
int
glfs_h_poll_cache_invalidation(struct glfs *fs, struct glfs_upcall *up_arg,
struct gf_upcall *upcall_data);
ssize_t
glfs_anonymous_preadv(struct glfs *fs, struct glfs_object *object,
const struct iovec *iovec, int iovcnt, off_t offset,
int flags);
ssize_t
glfs_anonymous_pwritev(struct glfs *fs, struct glfs_object *object,
const struct iovec *iovec, int iovcnt, off_t offset,
int flags);
struct glfs_object *
glfs_h_resolve_symlink(struct glfs *fs, struct glfs_object *object);
/* Deprecated structures that were passed to client applications, replaced by
* accessor functions. Do not use these in new applications, and update older
* usage.
*
* See http://review.gluster.org/14701 for more details.
*
* WARNING: These structures will be removed in the future.
*/
struct glfs_callback_arg {
struct glfs *fs;
enum glfs_upcall_reason reason;
void *event_arg;
};
struct glfs_callback_inode_arg {
struct glfs_object *object; /* Object which need to be acted upon */
int flags; /* Cache UPDATE/INVALIDATE flags */
struct stat buf; /* Latest stat of this entry */
unsigned int expire_time_attr; /* the amount of time for which
* the application need to cache
* this entry
*/
struct glfs_object *p_object; /* parent Object to be updated */
struct stat p_buf; /* Latest stat of parent dir handle */
struct glfs_object *oldp_object; /* Old parent Object
* to be updated */
struct stat oldp_buf; /* Latest stat of old parent
* dir handle */
};
struct dirent *
glfs_readdirbuf_get(struct glfs_fd *glfd);
gf_dirent_t *
glfd_entry_next(struct glfs_fd *glfd, int plus);
void
gf_dirent_to_dirent(gf_dirent_t *gf_dirent, struct dirent *dirent);
void
gf_lease_to_glfs_lease(struct gf_lease *gf_lease, struct glfs_lease *lease);
void
glfs_lease_to_gf_lease(struct glfs_lease *lease, struct gf_lease *gf_lease);
void
glfs_release_upcall(void *ptr);
int
get_fop_attr_glfd(dict_t **fop_attr, struct glfs_fd *glfd);
int
set_fop_attr_glfd(struct glfs_fd *glfd);
int
get_fop_attr_thrd_key(dict_t **fop_attr);
void
unset_fop_attr(dict_t **fop_attr);
/*
SYNOPSIS
glfs_statx: Fetch extended file attributes for the given path.
DESCRIPTION
This function fetches extended file attributes for the given path.
PARAMETERS
@fs: The 'virtual mount' object referencing a volume, under which file exists.
@path: Path of the file within the virtual mount.
@mask: Requested extended file attributes mask, (See mask defines above)
RETURN VALUES
-1 : Failure. @errno will be set with the type of failure.
0 : Filled in statxbuf with appropriate masks for valid items in the
structure.
ERRNO VALUES
EINVAL: fs is invalid
EINVAL: mask has unsupported bits set
Other errors as returned by stat(2)
*/
int
glfs_statx(struct glfs *fs, const char *path, unsigned int mask,
struct glfs_stat *statxbuf) GFAPI_PRIVATE(glfs_statx, future);
#endif /* !_GLFS_INTERNAL_H */
|