Lines Matching refs:ndd
40 static int validate_dimm(struct nvdimm_drvdata *ndd) in validate_dimm() argument
44 if (!ndd) in validate_dimm()
47 rc = nvdimm_check_config_data(ndd->dev); in validate_dimm()
49 dev_dbg(ndd->dev, "%ps: %s error: %d\n", in validate_dimm()
60 int nvdimm_init_nsarea(struct nvdimm_drvdata *ndd) in nvdimm_init_nsarea() argument
62 struct nd_cmd_get_config_size *cmd = &ndd->nsarea; in nvdimm_init_nsarea()
63 struct nvdimm_bus *nvdimm_bus = walk_to_nvdimm_bus(ndd->dev); in nvdimm_init_nsarea()
65 int rc = validate_dimm(ndd); in nvdimm_init_nsarea()
76 rc = nd_desc->ndctl(nd_desc, to_nvdimm(ndd->dev), in nvdimm_init_nsarea()
83 int nvdimm_get_config_data(struct nvdimm_drvdata *ndd, void *buf, in nvdimm_get_config_data() argument
86 struct nvdimm_bus *nvdimm_bus = walk_to_nvdimm_bus(ndd->dev); in nvdimm_get_config_data()
88 int rc = validate_dimm(ndd), cmd_rc = 0; in nvdimm_get_config_data()
95 if (offset + len > ndd->nsarea.config_size) in nvdimm_get_config_data()
98 max_cmd_size = min_t(u32, len, ndd->nsarea.max_xfer); in nvdimm_get_config_data()
112 rc = nd_desc->ndctl(nd_desc, to_nvdimm(ndd->dev), in nvdimm_get_config_data()
129 int nvdimm_set_config_data(struct nvdimm_drvdata *ndd, size_t offset, in nvdimm_set_config_data() argument
134 int rc = validate_dimm(ndd), cmd_rc = 0; in nvdimm_set_config_data()
135 struct nvdimm_bus *nvdimm_bus = walk_to_nvdimm_bus(ndd->dev); in nvdimm_set_config_data()
141 if (offset + len > ndd->nsarea.config_size) in nvdimm_set_config_data()
144 max_cmd_size = min_t(u32, len, ndd->nsarea.max_xfer); in nvdimm_set_config_data()
160 rc = nd_desc->ndctl(nd_desc, to_nvdimm(ndd->dev), in nvdimm_set_config_data()
224 struct nvdimm_drvdata *ndd = container_of(kref, typeof(*ndd), kref); in nvdimm_drvdata_release() local
225 struct device *dev = ndd->dev; in nvdimm_drvdata_release()
230 for_each_dpa_resource_safe(ndd, res, _r) in nvdimm_drvdata_release()
231 nvdimm_free_dpa(ndd, res); in nvdimm_drvdata_release()
234 kvfree(ndd->data); in nvdimm_drvdata_release()
235 kfree(ndd); in nvdimm_drvdata_release()
239 void get_ndd(struct nvdimm_drvdata *ndd) in get_ndd() argument
241 kref_get(&ndd->kref); in get_ndd()
244 void put_ndd(struct nvdimm_drvdata *ndd) in put_ndd() argument
246 if (ndd) in put_ndd()
247 kref_put(&ndd->kref, nvdimm_drvdata_release); in put_ndd()
319 static ssize_t __available_slots_show(struct nvdimm_drvdata *ndd, char *buf) in __available_slots_show() argument
325 if (!ndd) in __available_slots_show()
328 dev = ndd->dev; in __available_slots_show()
330 nfree = nd_label_nfree(ndd); in __available_slots_show()
734 struct nvdimm_drvdata *ndd = to_ndd(nd_mapping); in nd_pmem_max_contiguous_dpa() local
741 if (!ndd) in nd_pmem_max_contiguous_dpa()
748 nvdimm_bus = walk_to_nvdimm_bus(ndd->dev); in nd_pmem_max_contiguous_dpa()
751 for_each_dpa_resource(ndd, res) { in nd_pmem_max_contiguous_dpa()
781 struct nvdimm_drvdata *ndd = to_ndd(nd_mapping); in nd_pmem_available_dpa() local
786 if (!ndd) in nd_pmem_available_dpa()
795 for_each_dpa_resource(ndd, res) { in nd_pmem_available_dpa()
802 nd_dbg_dpa(nd_region, ndd, res, in nd_pmem_available_dpa()
820 void nvdimm_free_dpa(struct nvdimm_drvdata *ndd, struct resource *res) in nvdimm_free_dpa() argument
822 WARN_ON_ONCE(!is_nvdimm_bus_locked(ndd->dev)); in nvdimm_free_dpa()
824 __release_region(&ndd->dpa, res->start, resource_size(res)); in nvdimm_free_dpa()
827 struct resource *nvdimm_allocate_dpa(struct nvdimm_drvdata *ndd, in nvdimm_allocate_dpa() argument
837 WARN_ON_ONCE(!is_nvdimm_bus_locked(ndd->dev)); in nvdimm_allocate_dpa()
838 res = __request_region(&ndd->dpa, start, n, name, 0); in nvdimm_allocate_dpa()
851 resource_size_t nvdimm_allocated_dpa(struct nvdimm_drvdata *ndd, in nvdimm_allocated_dpa() argument
857 for_each_dpa_resource(ndd, res) in nvdimm_allocated_dpa()